var/home/core/zuul-output/0000755000175000017500000000000015071466665014545 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071504500015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004711443415071504471017711 0ustar rootrootOct 08 14:00:58 crc systemd[1]: Starting Kubernetes Kubelet... Oct 08 14:00:59 crc restorecon[4676]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 14:00:59 crc restorecon[4676]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 08 14:01:00 crc kubenswrapper[4789]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 14:01:00 crc kubenswrapper[4789]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 08 14:01:00 crc kubenswrapper[4789]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 14:01:00 crc kubenswrapper[4789]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 14:01:00 crc kubenswrapper[4789]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 08 14:01:00 crc kubenswrapper[4789]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.497013 4789 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502404 4789 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502421 4789 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502425 4789 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502429 4789 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502433 4789 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502437 4789 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502441 4789 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502445 4789 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502450 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502454 4789 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502458 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502462 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502466 4789 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502470 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502474 4789 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502477 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502481 4789 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502485 4789 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502488 4789 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502492 4789 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502495 4789 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502499 4789 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502502 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502507 4789 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502511 4789 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502514 4789 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502518 4789 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502521 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502525 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502530 4789 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502534 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502538 4789 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502543 4789 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502547 4789 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502552 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502556 4789 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502560 4789 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502563 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502567 4789 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502571 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502575 4789 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502579 4789 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502582 4789 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502586 4789 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502589 4789 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502593 4789 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502596 4789 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502600 4789 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502603 4789 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502606 4789 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502612 4789 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502616 4789 feature_gate.go:330] unrecognized feature gate: Example Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502621 4789 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502626 4789 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502630 4789 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502634 4789 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502637 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502641 4789 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502645 4789 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502648 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502652 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502655 4789 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502659 4789 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502662 4789 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502666 4789 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502669 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502674 4789 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502677 4789 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502681 4789 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502684 4789 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.502688 4789 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502761 4789 flags.go:64] FLAG: --address="0.0.0.0" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502770 4789 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502777 4789 flags.go:64] FLAG: --anonymous-auth="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502782 4789 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502788 4789 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502793 4789 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502799 4789 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502804 4789 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502808 4789 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502812 4789 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502816 4789 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502821 4789 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502825 4789 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502829 4789 flags.go:64] FLAG: --cgroup-root="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502833 4789 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502837 4789 flags.go:64] FLAG: --client-ca-file="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502841 4789 flags.go:64] FLAG: --cloud-config="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502844 4789 flags.go:64] FLAG: --cloud-provider="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502849 4789 flags.go:64] FLAG: --cluster-dns="[]" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502855 4789 flags.go:64] FLAG: --cluster-domain="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502859 4789 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502863 4789 flags.go:64] FLAG: --config-dir="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502867 4789 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502871 4789 flags.go:64] FLAG: --container-log-max-files="5" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502877 4789 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502880 4789 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502884 4789 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502889 4789 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502892 4789 flags.go:64] FLAG: --contention-profiling="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502896 4789 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502900 4789 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502905 4789 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502909 4789 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502914 4789 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502919 4789 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502923 4789 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502928 4789 flags.go:64] FLAG: --enable-load-reader="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502932 4789 flags.go:64] FLAG: --enable-server="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502936 4789 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502941 4789 flags.go:64] FLAG: --event-burst="100" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502945 4789 flags.go:64] FLAG: --event-qps="50" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502950 4789 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502954 4789 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502958 4789 flags.go:64] FLAG: --eviction-hard="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502963 4789 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502967 4789 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502972 4789 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502976 4789 flags.go:64] FLAG: --eviction-soft="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502980 4789 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.502983 4789 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503001 4789 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503005 4789 flags.go:64] FLAG: --experimental-mounter-path="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503009 4789 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503013 4789 flags.go:64] FLAG: --fail-swap-on="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503017 4789 flags.go:64] FLAG: --feature-gates="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503022 4789 flags.go:64] FLAG: --file-check-frequency="20s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503026 4789 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503030 4789 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503035 4789 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503039 4789 flags.go:64] FLAG: --healthz-port="10248" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503044 4789 flags.go:64] FLAG: --help="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503048 4789 flags.go:64] FLAG: --hostname-override="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503052 4789 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503056 4789 flags.go:64] FLAG: --http-check-frequency="20s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503060 4789 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503064 4789 flags.go:64] FLAG: --image-credential-provider-config="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503068 4789 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503072 4789 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503076 4789 flags.go:64] FLAG: --image-service-endpoint="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503080 4789 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503084 4789 flags.go:64] FLAG: --kube-api-burst="100" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503088 4789 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503092 4789 flags.go:64] FLAG: --kube-api-qps="50" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503096 4789 flags.go:64] FLAG: --kube-reserved="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503100 4789 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503104 4789 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503108 4789 flags.go:64] FLAG: --kubelet-cgroups="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503112 4789 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503117 4789 flags.go:64] FLAG: --lock-file="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503121 4789 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503125 4789 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503129 4789 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503135 4789 flags.go:64] FLAG: --log-json-split-stream="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503139 4789 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503143 4789 flags.go:64] FLAG: --log-text-split-stream="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503147 4789 flags.go:64] FLAG: --logging-format="text" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503151 4789 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503155 4789 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503159 4789 flags.go:64] FLAG: --manifest-url="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503163 4789 flags.go:64] FLAG: --manifest-url-header="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503171 4789 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503175 4789 flags.go:64] FLAG: --max-open-files="1000000" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503181 4789 flags.go:64] FLAG: --max-pods="110" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503185 4789 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503189 4789 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503193 4789 flags.go:64] FLAG: --memory-manager-policy="None" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503197 4789 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503202 4789 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503206 4789 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503210 4789 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503220 4789 flags.go:64] FLAG: --node-status-max-images="50" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503224 4789 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503228 4789 flags.go:64] FLAG: --oom-score-adj="-999" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503232 4789 flags.go:64] FLAG: --pod-cidr="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503237 4789 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503244 4789 flags.go:64] FLAG: --pod-manifest-path="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503248 4789 flags.go:64] FLAG: --pod-max-pids="-1" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503252 4789 flags.go:64] FLAG: --pods-per-core="0" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503256 4789 flags.go:64] FLAG: --port="10250" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503260 4789 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503265 4789 flags.go:64] FLAG: --provider-id="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503268 4789 flags.go:64] FLAG: --qos-reserved="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503273 4789 flags.go:64] FLAG: --read-only-port="10255" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503277 4789 flags.go:64] FLAG: --register-node="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503281 4789 flags.go:64] FLAG: --register-schedulable="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503285 4789 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503293 4789 flags.go:64] FLAG: --registry-burst="10" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503297 4789 flags.go:64] FLAG: --registry-qps="5" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503301 4789 flags.go:64] FLAG: --reserved-cpus="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503304 4789 flags.go:64] FLAG: --reserved-memory="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503310 4789 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503314 4789 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503318 4789 flags.go:64] FLAG: --rotate-certificates="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503322 4789 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503326 4789 flags.go:64] FLAG: --runonce="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503330 4789 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503334 4789 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503338 4789 flags.go:64] FLAG: --seccomp-default="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503342 4789 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503346 4789 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503350 4789 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503354 4789 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503358 4789 flags.go:64] FLAG: --storage-driver-password="root" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503363 4789 flags.go:64] FLAG: --storage-driver-secure="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503367 4789 flags.go:64] FLAG: --storage-driver-table="stats" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503371 4789 flags.go:64] FLAG: --storage-driver-user="root" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503375 4789 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503379 4789 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503383 4789 flags.go:64] FLAG: --system-cgroups="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503387 4789 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503394 4789 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503397 4789 flags.go:64] FLAG: --tls-cert-file="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503402 4789 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503408 4789 flags.go:64] FLAG: --tls-min-version="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503412 4789 flags.go:64] FLAG: --tls-private-key-file="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503416 4789 flags.go:64] FLAG: --topology-manager-policy="none" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503420 4789 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503424 4789 flags.go:64] FLAG: --topology-manager-scope="container" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503428 4789 flags.go:64] FLAG: --v="2" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503435 4789 flags.go:64] FLAG: --version="false" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503440 4789 flags.go:64] FLAG: --vmodule="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503445 4789 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503449 4789 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503544 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503550 4789 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503554 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503558 4789 feature_gate.go:330] unrecognized feature gate: Example Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503562 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503566 4789 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503569 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503573 4789 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503576 4789 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503579 4789 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503632 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503636 4789 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503641 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503644 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503666 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503669 4789 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503673 4789 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503677 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503681 4789 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503684 4789 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503688 4789 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503692 4789 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503695 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503699 4789 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503703 4789 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503708 4789 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503712 4789 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503716 4789 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503720 4789 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503724 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503727 4789 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503731 4789 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503736 4789 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503740 4789 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503744 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503748 4789 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503752 4789 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503756 4789 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503759 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503763 4789 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503766 4789 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503770 4789 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503773 4789 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503777 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503781 4789 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503786 4789 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503812 4789 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503818 4789 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503822 4789 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503827 4789 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503831 4789 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503835 4789 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503839 4789 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503843 4789 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503847 4789 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503850 4789 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503857 4789 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503860 4789 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503864 4789 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503867 4789 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503871 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503874 4789 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503879 4789 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503882 4789 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503887 4789 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503891 4789 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503895 4789 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503898 4789 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503903 4789 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503906 4789 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.503910 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.503922 4789 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.513375 4789 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.513440 4789 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513532 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513544 4789 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513550 4789 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513557 4789 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513564 4789 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513571 4789 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513577 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513583 4789 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513589 4789 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513596 4789 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513604 4789 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513614 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513620 4789 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513627 4789 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513633 4789 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513640 4789 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513649 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513656 4789 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513664 4789 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513673 4789 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513681 4789 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513688 4789 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513697 4789 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513704 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513712 4789 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513720 4789 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513727 4789 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513734 4789 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513740 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513747 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513752 4789 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513758 4789 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513764 4789 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513769 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513776 4789 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513781 4789 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513787 4789 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513793 4789 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513798 4789 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513804 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513809 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513815 4789 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513820 4789 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513826 4789 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513831 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513837 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513842 4789 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513847 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513852 4789 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513857 4789 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513863 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513868 4789 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513874 4789 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513879 4789 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513884 4789 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513891 4789 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513896 4789 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513902 4789 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513907 4789 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513912 4789 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513917 4789 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513922 4789 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513928 4789 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513933 4789 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513938 4789 feature_gate.go:330] unrecognized feature gate: Example Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513943 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513948 4789 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513953 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513958 4789 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.513964 4789 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514032 4789 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.514046 4789 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514261 4789 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514271 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514276 4789 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514282 4789 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514286 4789 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514291 4789 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514296 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514302 4789 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514307 4789 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514313 4789 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514318 4789 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514324 4789 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514330 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514335 4789 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514340 4789 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514345 4789 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514350 4789 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514357 4789 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514365 4789 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514371 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514376 4789 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514382 4789 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514390 4789 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514398 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514405 4789 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514412 4789 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514418 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514423 4789 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514428 4789 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514433 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514438 4789 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514446 4789 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514451 4789 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514456 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514470 4789 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514476 4789 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514481 4789 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514486 4789 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514490 4789 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514496 4789 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514503 4789 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514509 4789 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514514 4789 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514519 4789 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514524 4789 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514531 4789 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514537 4789 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514543 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514548 4789 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514553 4789 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514559 4789 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514565 4789 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514570 4789 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514575 4789 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514580 4789 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514584 4789 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514590 4789 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514595 4789 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514599 4789 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514604 4789 feature_gate.go:330] unrecognized feature gate: Example Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514609 4789 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514614 4789 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514618 4789 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514623 4789 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514628 4789 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514633 4789 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514638 4789 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514644 4789 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514648 4789 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514653 4789 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.514666 4789 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.514673 4789 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.515922 4789 server.go:940] "Client rotation is on, will bootstrap in background" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.526081 4789 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.526223 4789 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.527794 4789 server.go:997] "Starting client certificate rotation" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.527829 4789 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.528961 4789 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-10 01:59:04.932220614 +0000 UTC Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.529082 4789 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1499h58m4.403141916s for next certificate rotation Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.559490 4789 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.561448 4789 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.582808 4789 log.go:25] "Validated CRI v1 runtime API" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.616610 4789 log.go:25] "Validated CRI v1 image API" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.619375 4789 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.624682 4789 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-08-13-55-57-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.624767 4789 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.653952 4789 manager.go:217] Machine: {Timestamp:2025-10-08 14:01:00.651292313 +0000 UTC m=+0.558039885 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3aa13740-a16f-4d58-8132-6848dda41690 BootID:59598eb2-dd3c-404b-b870-d7b0c91c1ae8 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:db:19:ae Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:db:19:ae Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:99:da:91 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ad:83:6d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:76:3a:ab Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:1f:1e:ed Speed:-1 Mtu:1496} {Name:eth10 MacAddress:1a:89:10:27:78:cf Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:52:b9:63:7a:54:42 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.654717 4789 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.654954 4789 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.657152 4789 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.657518 4789 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.657569 4789 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.657930 4789 topology_manager.go:138] "Creating topology manager with none policy" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.657953 4789 container_manager_linux.go:303] "Creating device plugin manager" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.658599 4789 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.658651 4789 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.658960 4789 state_mem.go:36] "Initialized new in-memory state store" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.659148 4789 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.662619 4789 kubelet.go:418] "Attempting to sync node with API server" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.662645 4789 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.662663 4789 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.662678 4789 kubelet.go:324] "Adding apiserver pod source" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.662695 4789 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.672636 4789 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.674248 4789 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.678031 4789 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.679334 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.679439 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679500 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679520 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679528 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679534 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679545 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679552 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679559 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679570 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679579 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679587 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.679528 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679627 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.679636 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.679658 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.680570 4789 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.681117 4789 server.go:1280] "Started kubelet" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.681217 4789 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.681883 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.681960 4789 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.682578 4789 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.682608 4789 server.go:460] "Adding debug handlers to kubelet server" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.683138 4789 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.683186 4789 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.683308 4789 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 15:00:52.547106323 +0000 UTC Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.683357 4789 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1296h59m51.863752048s for next certificate rotation Oct 08 14:01:00 crc systemd[1]: Started Kubernetes Kubelet. Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.683445 4789 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.683471 4789 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.683599 4789 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.683782 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.683908 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="200ms" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.685409 4789 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.685426 4789 factory.go:55] Registering systemd factory Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.685438 4789 factory.go:221] Registration of the systemd container factory successfully Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.685573 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.685657 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.685747 4789 factory.go:153] Registering CRI-O factory Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.685768 4789 factory.go:221] Registration of the crio container factory successfully Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.685798 4789 factory.go:103] Registering Raw factory Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.685814 4789 manager.go:1196] Started watching for new ooms in manager Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.686957 4789 manager.go:319] Starting recovery of all containers Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.691513 4789 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.214:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c88e279b79ff8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 14:01:00.68108492 +0000 UTC m=+0.587832412,LastTimestamp:2025-10-08 14:01:00.68108492 +0000 UTC m=+0.587832412,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698682 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698755 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698766 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698775 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698787 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698796 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698806 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698817 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698831 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698844 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698857 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698867 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698876 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698891 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698900 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698910 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698941 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698952 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698961 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698972 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.698999 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699010 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699020 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699029 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699039 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699048 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699060 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699070 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699080 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699095 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699123 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699153 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699164 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699177 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699190 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699203 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699213 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699226 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699236 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699246 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699259 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699269 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699280 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699291 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699301 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699312 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699322 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699332 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699345 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699356 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699368 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699378 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699394 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699431 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699448 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699459 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699471 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699479 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699488 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699498 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699508 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699518 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699526 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699536 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699546 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699555 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699566 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699575 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699584 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699594 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699604 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699613 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699623 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699633 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699641 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699650 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699662 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699673 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699682 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699691 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699701 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699712 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699721 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699731 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699742 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699754 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699765 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699777 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699787 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699798 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699807 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699818 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699828 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699839 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699850 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699860 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699870 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699879 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699891 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699901 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699911 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699923 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699935 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699945 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699967 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.699999 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700014 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700026 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700037 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700049 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700060 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700070 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700081 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700090 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700100 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700109 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700164 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700175 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700183 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700192 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700201 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700216 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700225 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700234 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700242 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700259 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700269 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700279 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700289 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700300 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700312 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700321 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700331 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700342 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700352 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700361 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700374 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700385 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700395 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700406 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700418 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700457 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700469 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700481 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700498 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700509 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700521 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700533 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700543 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700554 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700566 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700577 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700589 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700601 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700612 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700622 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700632 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700645 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700654 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700666 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700677 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700689 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700715 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.700727 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704172 4789 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704225 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704248 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704263 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704277 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704299 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704313 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704326 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704337 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704351 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704362 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704375 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704389 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704403 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704416 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704428 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704442 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704454 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704467 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704481 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704492 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704505 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704517 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704531 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704542 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704556 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704567 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704579 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704591 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704604 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704615 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704627 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704638 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704652 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704663 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704674 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704687 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704725 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704738 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704750 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704762 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704778 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704794 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704810 4789 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704826 4789 reconstruct.go:97] "Volume reconstruction finished" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.704836 4789 reconciler.go:26] "Reconciler: start to sync state" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.711112 4789 manager.go:324] Recovery completed Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.720863 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.722494 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.722534 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.722547 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.723953 4789 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.723971 4789 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.724034 4789 state_mem.go:36] "Initialized new in-memory state store" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.727089 4789 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.728605 4789 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.728653 4789 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.728718 4789 kubelet.go:2335] "Starting kubelet main sync loop" Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.728868 4789 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 08 14:01:00 crc kubenswrapper[4789]: W1008 14:01:00.730452 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.730526 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.746810 4789 policy_none.go:49] "None policy: Start" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.747673 4789 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.747801 4789 state_mem.go:35] "Initializing new in-memory state store" Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.784213 4789 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.803935 4789 manager.go:334] "Starting Device Plugin manager" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.804190 4789 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.804211 4789 server.go:79] "Starting device plugin registration server" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.804638 4789 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.804689 4789 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.804875 4789 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.805038 4789 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.805049 4789 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.828910 4789 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.828960 4789 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.829062 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.830562 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.830609 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.830621 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.830854 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.831686 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.831736 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.832730 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.832807 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.832827 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.832844 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.832858 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.832878 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.833631 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.833729 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.833785 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.834580 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.834604 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.834616 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.834730 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.834848 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.834876 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.834886 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.835137 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.835173 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.835474 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.835521 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.835538 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.835736 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.835812 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.835835 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.835967 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.836014 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.836027 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.836559 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.836583 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.836592 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.836766 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.836820 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.836839 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.837117 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.837169 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.837960 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.838003 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.838013 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.885076 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="400ms" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.905152 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906127 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906220 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906259 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906300 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906341 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906383 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906417 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906451 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906491 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906529 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906560 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906595 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906634 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906743 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906787 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906860 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906899 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906910 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:00 crc kubenswrapper[4789]: I1008 14:01:00.906968 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 14:01:00 crc kubenswrapper[4789]: E1008 14:01:00.907484 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.007947 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008070 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008112 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008156 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008201 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008248 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008270 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008289 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008214 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008307 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008387 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008417 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008435 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008457 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008475 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008491 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008509 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008513 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008512 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008530 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008588 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008608 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008610 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008631 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008639 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008592 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008605 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008714 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008640 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.008729 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.108285 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.110271 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.110337 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.110356 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.110396 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 14:01:01 crc kubenswrapper[4789]: E1008 14:01:01.110968 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.171671 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.176724 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.192026 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.216812 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.224220 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:01 crc kubenswrapper[4789]: W1008 14:01:01.226012 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-0d3e5727a37d4e1c8545ca1524cc64a8a0623236b75690f3876618de8b82c44a WatchSource:0}: Error finding container 0d3e5727a37d4e1c8545ca1524cc64a8a0623236b75690f3876618de8b82c44a: Status 404 returned error can't find the container with id 0d3e5727a37d4e1c8545ca1524cc64a8a0623236b75690f3876618de8b82c44a Oct 08 14:01:01 crc kubenswrapper[4789]: W1008 14:01:01.227376 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-d287f6ca5f7af98b50ddcd2058a0b8e876908def636a1a6c9b6ab460c5d35fbb WatchSource:0}: Error finding container d287f6ca5f7af98b50ddcd2058a0b8e876908def636a1a6c9b6ab460c5d35fbb: Status 404 returned error can't find the container with id d287f6ca5f7af98b50ddcd2058a0b8e876908def636a1a6c9b6ab460c5d35fbb Oct 08 14:01:01 crc kubenswrapper[4789]: W1008 14:01:01.234422 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-2a3c7ce5eac4df8d19aad419b4edc271a32a58e026f402d6d09e625a7253274c WatchSource:0}: Error finding container 2a3c7ce5eac4df8d19aad419b4edc271a32a58e026f402d6d09e625a7253274c: Status 404 returned error can't find the container with id 2a3c7ce5eac4df8d19aad419b4edc271a32a58e026f402d6d09e625a7253274c Oct 08 14:01:01 crc kubenswrapper[4789]: W1008 14:01:01.242766 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f350beeb91b97ce2ecaff049a07c4f69cd9fee7760476eef2d576e03757da64b WatchSource:0}: Error finding container f350beeb91b97ce2ecaff049a07c4f69cd9fee7760476eef2d576e03757da64b: Status 404 returned error can't find the container with id f350beeb91b97ce2ecaff049a07c4f69cd9fee7760476eef2d576e03757da64b Oct 08 14:01:01 crc kubenswrapper[4789]: W1008 14:01:01.251513 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-6d55db71775c8b24255e1b86277ca210ae834b5c5d41e2da10137c9b6ac0d71b WatchSource:0}: Error finding container 6d55db71775c8b24255e1b86277ca210ae834b5c5d41e2da10137c9b6ac0d71b: Status 404 returned error can't find the container with id 6d55db71775c8b24255e1b86277ca210ae834b5c5d41e2da10137c9b6ac0d71b Oct 08 14:01:01 crc kubenswrapper[4789]: E1008 14:01:01.286035 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="800ms" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.511130 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.513120 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.513434 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.513447 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.513482 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 14:01:01 crc kubenswrapper[4789]: E1008 14:01:01.514056 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 08 14:01:01 crc kubenswrapper[4789]: W1008 14:01:01.605488 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:01 crc kubenswrapper[4789]: E1008 14:01:01.605588 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 08 14:01:01 crc kubenswrapper[4789]: W1008 14:01:01.674111 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:01 crc kubenswrapper[4789]: E1008 14:01:01.674186 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.683788 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.733711 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d287f6ca5f7af98b50ddcd2058a0b8e876908def636a1a6c9b6ab460c5d35fbb"} Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.735001 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0d3e5727a37d4e1c8545ca1524cc64a8a0623236b75690f3876618de8b82c44a"} Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.736545 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6d55db71775c8b24255e1b86277ca210ae834b5c5d41e2da10137c9b6ac0d71b"} Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.738032 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f350beeb91b97ce2ecaff049a07c4f69cd9fee7760476eef2d576e03757da64b"} Oct 08 14:01:01 crc kubenswrapper[4789]: I1008 14:01:01.739050 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2a3c7ce5eac4df8d19aad419b4edc271a32a58e026f402d6d09e625a7253274c"} Oct 08 14:01:01 crc kubenswrapper[4789]: W1008 14:01:01.899243 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:01 crc kubenswrapper[4789]: E1008 14:01:01.899337 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 08 14:01:01 crc kubenswrapper[4789]: W1008 14:01:01.998723 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:01 crc kubenswrapper[4789]: E1008 14:01:01.998854 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 08 14:01:02 crc kubenswrapper[4789]: E1008 14:01:02.087188 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="1.6s" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.315077 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.316359 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.316403 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.316416 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.316446 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 14:01:02 crc kubenswrapper[4789]: E1008 14:01:02.316914 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.683358 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.743388 4789 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df" exitCode=0 Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.743478 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df"} Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.743518 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.744792 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.744826 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.744839 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.746387 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"32e9e9c47612a7965e2e458b61123ae22bcbfabb102878302e2962ba55b8370a"} Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.746414 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"175ebc8a5b7234d4118cd6c318781277ab2e18cce1cc44f6d07d70e57e9c5281"} Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.749637 4789 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9" exitCode=0 Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.749691 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9"} Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.749796 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.750874 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.750949 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.750975 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.751931 4789 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3de4f2327dd58e0ebfedc75f4418f1ebe6be8276a54d5dd16bd07a8611afd06e" exitCode=0 Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.752014 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3de4f2327dd58e0ebfedc75f4418f1ebe6be8276a54d5dd16bd07a8611afd06e"} Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.752039 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.752766 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.752800 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.752813 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.753505 4789 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4" exitCode=0 Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.753535 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4"} Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.753588 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.757264 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.757826 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.757844 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.759953 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.760849 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.760880 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:02 crc kubenswrapper[4789]: I1008 14:01:02.760896 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.683665 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:03 crc kubenswrapper[4789]: E1008 14:01:03.688745 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="3.2s" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.758223 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c34f5f170b7d878df231057281140e6881509ff3856c29d13d935c783143d99f"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.758264 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"584b686828d861a74791e8edd92fba9a6b5a0418a476d08badc734b6f80d1b06"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.758346 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.760623 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.760678 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.760691 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:03 crc kubenswrapper[4789]: W1008 14:01:03.766911 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Oct 08 14:01:03 crc kubenswrapper[4789]: E1008 14:01:03.767039 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.770476 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.770525 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.770539 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.770551 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.772358 4789 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="eb96e4e7eaafdfb66edf48923911377c6fcba80f431604297d548e6cf14da1ed" exitCode=0 Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.772418 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"eb96e4e7eaafdfb66edf48923911377c6fcba80f431604297d548e6cf14da1ed"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.772480 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.773528 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.773557 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.773570 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.773906 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fcb4658ef6d767e0b260c91435cffde1ce49a2ef282408f85c68c7b459323f4a"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.774008 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.775860 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.775882 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.775893 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.779475 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.779508 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.779523 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a"} Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.779580 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.780502 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.780539 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.780557 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.917478 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.918691 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.918726 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.918737 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:03 crc kubenswrapper[4789]: I1008 14:01:03.918764 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 14:01:03 crc kubenswrapper[4789]: E1008 14:01:03.919336 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.786921 4789 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f90f7f70eaf5033b84d7a9029a1197da1bd1f0612286aeabf3d218340a799668" exitCode=0 Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.787057 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f90f7f70eaf5033b84d7a9029a1197da1bd1f0612286aeabf3d218340a799668"} Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.787241 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.788924 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.789036 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.789061 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.794433 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221"} Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.794546 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.794586 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.794627 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.794704 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.794640 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.796520 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.796559 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.796571 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.796703 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.796761 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.796782 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.797157 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.797199 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.797256 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.797215 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.797280 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:04 crc kubenswrapper[4789]: I1008 14:01:04.797303 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:05 crc kubenswrapper[4789]: I1008 14:01:05.804033 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"37efd9b75e8e61b12b78efc71fed7cd9763dec6aab84ce3a2750a37a1c604b3d"} Oct 08 14:01:05 crc kubenswrapper[4789]: I1008 14:01:05.804124 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fc6975a812860b760a75b09aed3ebbe89e2453b297b7cf4c0361791919cfa579"} Oct 08 14:01:05 crc kubenswrapper[4789]: I1008 14:01:05.804145 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:05 crc kubenswrapper[4789]: I1008 14:01:05.804148 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"21cdb40a34fb8828f2c9a1f182102e766a3d6bbfd08f2336166dc3cc847a9cd3"} Oct 08 14:01:05 crc kubenswrapper[4789]: I1008 14:01:05.804239 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:05 crc kubenswrapper[4789]: I1008 14:01:05.805405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:05 crc kubenswrapper[4789]: I1008 14:01:05.805449 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:05 crc kubenswrapper[4789]: I1008 14:01:05.805462 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.087642 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.087879 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.089392 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.089443 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.089460 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.635830 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.636207 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.638156 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.638219 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.638239 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.647983 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.813903 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1674c2135065bf6fc3e5b9ea2beb38b184683ae07867d7a3106389fef75b2241"} Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.814090 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.814165 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.814099 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f9c088df3700b61b3e339802661a13df97da812937f72fda3d6182dff0490a13"} Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.814447 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.815917 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.815945 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.815973 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.815980 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.816030 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.816015 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.816229 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.816279 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:06 crc kubenswrapper[4789]: I1008 14:01:06.816305 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.120060 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.121560 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.121632 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.121666 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.121711 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.763719 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.816927 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.816935 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.818569 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.818614 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.818638 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.819108 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.819172 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:07 crc kubenswrapper[4789]: I1008 14:01:07.819193 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.400224 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.819639 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.821135 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.821188 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.821208 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.858845 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.859154 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.865405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.865527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:08 crc kubenswrapper[4789]: I1008 14:01:08.865558 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:09 crc kubenswrapper[4789]: I1008 14:01:09.668610 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 08 14:01:09 crc kubenswrapper[4789]: I1008 14:01:09.822310 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:09 crc kubenswrapper[4789]: I1008 14:01:09.823376 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:09 crc kubenswrapper[4789]: I1008 14:01:09.823455 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:09 crc kubenswrapper[4789]: I1008 14:01:09.823478 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.270219 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.270484 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.270553 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.272183 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.272234 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.272252 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.620900 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.695447 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.824766 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.825876 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.825918 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:10 crc kubenswrapper[4789]: I1008 14:01:10.825931 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:10 crc kubenswrapper[4789]: E1008 14:01:10.829319 4789 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 14:01:11 crc kubenswrapper[4789]: I1008 14:01:11.827072 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:11 crc kubenswrapper[4789]: I1008 14:01:11.828550 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:11 crc kubenswrapper[4789]: I1008 14:01:11.828744 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:11 crc kubenswrapper[4789]: I1008 14:01:11.828767 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:11 crc kubenswrapper[4789]: I1008 14:01:11.833675 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:12 crc kubenswrapper[4789]: I1008 14:01:12.829444 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:12 crc kubenswrapper[4789]: I1008 14:01:12.830681 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:12 crc kubenswrapper[4789]: I1008 14:01:12.830746 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:12 crc kubenswrapper[4789]: I1008 14:01:12.830773 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:13 crc kubenswrapper[4789]: I1008 14:01:13.270579 4789 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 14:01:13 crc kubenswrapper[4789]: I1008 14:01:13.270723 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 14:01:13 crc kubenswrapper[4789]: I1008 14:01:13.746898 4789 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 08 14:01:13 crc kubenswrapper[4789]: I1008 14:01:13.747403 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 08 14:01:14 crc kubenswrapper[4789]: W1008 14:01:14.256927 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 08 14:01:14 crc kubenswrapper[4789]: I1008 14:01:14.257095 4789 trace.go:236] Trace[2087345933]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 14:01:04.255) (total time: 10001ms): Oct 08 14:01:14 crc kubenswrapper[4789]: Trace[2087345933]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:01:14.256) Oct 08 14:01:14 crc kubenswrapper[4789]: Trace[2087345933]: [10.001320063s] [10.001320063s] END Oct 08 14:01:14 crc kubenswrapper[4789]: E1008 14:01:14.257135 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 08 14:01:14 crc kubenswrapper[4789]: W1008 14:01:14.498712 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 08 14:01:14 crc kubenswrapper[4789]: I1008 14:01:14.498822 4789 trace.go:236] Trace[26143234]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 14:01:04.497) (total time: 10000ms): Oct 08 14:01:14 crc kubenswrapper[4789]: Trace[26143234]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (14:01:14.498) Oct 08 14:01:14 crc kubenswrapper[4789]: Trace[26143234]: [10.000956444s] [10.000956444s] END Oct 08 14:01:14 crc kubenswrapper[4789]: E1008 14:01:14.498850 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 08 14:01:14 crc kubenswrapper[4789]: I1008 14:01:14.684381 4789 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 08 14:01:14 crc kubenswrapper[4789]: W1008 14:01:14.724185 4789 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 08 14:01:14 crc kubenswrapper[4789]: I1008 14:01:14.724355 4789 trace.go:236] Trace[360497887]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 14:01:04.722) (total time: 10002ms): Oct 08 14:01:14 crc kubenswrapper[4789]: Trace[360497887]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:01:14.724) Oct 08 14:01:14 crc kubenswrapper[4789]: Trace[360497887]: [10.00202411s] [10.00202411s] END Oct 08 14:01:14 crc kubenswrapper[4789]: E1008 14:01:14.724400 4789 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 08 14:01:14 crc kubenswrapper[4789]: I1008 14:01:14.956584 4789 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 14:01:14 crc kubenswrapper[4789]: I1008 14:01:14.956688 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 14:01:14 crc kubenswrapper[4789]: I1008 14:01:14.978698 4789 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]log ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]etcd ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/generic-apiserver-start-informers ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/priority-and-fairness-filter ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/start-apiextensions-informers ok Oct 08 14:01:14 crc kubenswrapper[4789]: [-]poststarthook/start-apiextensions-controllers failed: reason withheld Oct 08 14:01:14 crc kubenswrapper[4789]: [-]poststarthook/crd-informer-synced failed: reason withheld Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/start-system-namespaces-controller ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 08 14:01:14 crc kubenswrapper[4789]: [-]poststarthook/start-service-ip-repair-controllers failed: reason withheld Oct 08 14:01:14 crc kubenswrapper[4789]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 08 14:01:14 crc kubenswrapper[4789]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 08 14:01:14 crc kubenswrapper[4789]: [-]poststarthook/priority-and-fairness-config-producer failed: reason withheld Oct 08 14:01:14 crc kubenswrapper[4789]: [-]poststarthook/bootstrap-controller failed: reason withheld Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/start-kube-aggregator-informers ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 08 14:01:14 crc kubenswrapper[4789]: [-]poststarthook/apiservice-registration-controller failed: reason withheld Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 08 14:01:14 crc kubenswrapper[4789]: [-]poststarthook/apiservice-discovery-controller failed: reason withheld Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]autoregister-completion ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/apiservice-openapi-controller ok Oct 08 14:01:14 crc kubenswrapper[4789]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 08 14:01:14 crc kubenswrapper[4789]: livez check failed Oct 08 14:01:14 crc kubenswrapper[4789]: I1008 14:01:14.978814 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 14:01:17 crc kubenswrapper[4789]: I1008 14:01:17.819548 4789 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.405090 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.405258 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.406566 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.406600 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.406610 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.410137 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.843572 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.844501 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.844551 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.844564 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.892894 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.893133 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.894338 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.894379 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.894392 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:18 crc kubenswrapper[4789]: I1008 14:01:18.908356 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 08 14:01:19 crc kubenswrapper[4789]: I1008 14:01:19.700617 4789 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 08 14:01:19 crc kubenswrapper[4789]: I1008 14:01:19.845839 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:19 crc kubenswrapper[4789]: I1008 14:01:19.846568 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:19 crc kubenswrapper[4789]: I1008 14:01:19.846592 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:19 crc kubenswrapper[4789]: I1008 14:01:19.846603 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:19 crc kubenswrapper[4789]: E1008 14:01:19.952356 4789 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 08 14:01:19 crc kubenswrapper[4789]: I1008 14:01:19.953787 4789 trace.go:236] Trace[1142424044]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 14:01:08.710) (total time: 11243ms): Oct 08 14:01:19 crc kubenswrapper[4789]: Trace[1142424044]: ---"Objects listed" error: 11243ms (14:01:19.953) Oct 08 14:01:19 crc kubenswrapper[4789]: Trace[1142424044]: [11.243433111s] [11.243433111s] END Oct 08 14:01:19 crc kubenswrapper[4789]: I1008 14:01:19.953814 4789 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 08 14:01:19 crc kubenswrapper[4789]: I1008 14:01:19.957626 4789 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 08 14:01:19 crc kubenswrapper[4789]: E1008 14:01:19.958620 4789 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.041593 4789 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.041652 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.044534 4789 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:47662->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.044600 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:47662->192.168.126.11:17697: read: connection reset by peer" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.655602 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.655759 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.656738 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.656761 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.656770 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.659759 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 14:01:20 crc kubenswrapper[4789]: E1008 14:01:20.829768 4789 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.849305 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.850854 4789 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221" exitCode=255 Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.850963 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221"} Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.851008 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.851188 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.852025 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.852061 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.852089 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.852241 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.852273 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.852283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.852841 4789 scope.go:117] "RemoveContainer" containerID="099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221" Oct 08 14:01:20 crc kubenswrapper[4789]: I1008 14:01:20.958544 4789 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.676716 4789 apiserver.go:52] "Watching apiserver" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.679368 4789 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.679804 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-r858x","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.680262 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.680332 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.680464 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.680496 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.680394 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.680560 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.680636 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.680853 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.680883 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-r858x" Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.680892 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.682677 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.682751 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.683642 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.683703 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.683728 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.684258 4789 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.684487 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.684596 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.684922 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.684953 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.685005 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.685937 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.686276 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.707232 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.718072 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.719843 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-4m4jj"] Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.720703 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.723151 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.723449 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-94gx2"] Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.723564 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.723594 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.723822 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-5428v"] Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.724016 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.724016 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.724684 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.724726 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.730577 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.730663 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.731895 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.732480 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.732518 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.732553 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.735273 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.738945 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.751720 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.768201 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770279 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770327 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770349 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770368 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770388 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770405 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770419 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770433 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770452 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770471 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770512 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770528 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770543 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770557 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770571 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770586 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770608 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770624 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770640 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770656 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770672 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770690 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770706 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770721 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770737 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770757 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770773 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770789 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770805 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770820 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770849 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770879 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770898 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770914 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770933 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770952 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.770967 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771000 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771018 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771034 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771054 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771070 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771087 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771105 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771227 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771246 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771262 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771281 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771300 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771316 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771331 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771347 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771364 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771397 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771415 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771431 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771450 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771465 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771484 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771503 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771518 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771534 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771567 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771583 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771602 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771618 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771634 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771654 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771684 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771704 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771724 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771743 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771729 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771765 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771783 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771798 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771814 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771832 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771847 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771866 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771881 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771900 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771917 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771931 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771947 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771963 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771980 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.771982 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772021 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772037 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772055 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772070 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772086 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772102 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772118 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772135 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772154 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772173 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772196 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772222 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772253 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772269 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772287 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772315 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772333 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772350 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772367 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772384 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772401 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772419 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772437 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772453 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772470 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772504 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772523 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772541 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772557 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772576 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772592 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772607 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772623 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772639 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772655 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772670 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772687 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772704 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772748 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772765 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772782 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772800 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772819 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772836 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772852 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772868 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772886 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772906 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772925 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772945 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772963 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772996 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773023 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773043 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773064 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773087 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773107 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773127 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773147 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773166 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773184 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773202 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773220 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773240 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773259 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773280 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773299 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773316 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773334 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773351 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773367 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773389 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773405 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773422 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773444 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773464 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773486 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773511 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773531 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773552 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773571 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773590 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773611 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773631 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773654 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773677 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773696 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773717 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773737 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773758 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773779 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773799 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773814 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773831 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773846 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773862 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773879 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773897 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773916 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773933 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773951 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773969 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774003 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774022 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774042 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774062 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774082 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774098 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774116 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774164 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774186 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774208 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774227 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774246 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774267 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774284 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774302 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5bxb\" (UniqueName: \"kubernetes.io/projected/de6093ed-9813-429d-812d-26bddadd2fa6-kube-api-access-x5bxb\") pod \"node-resolver-r858x\" (UID: \"de6093ed-9813-429d-812d-26bddadd2fa6\") " pod="openshift-dns/node-resolver-r858x" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774321 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/de6093ed-9813-429d-812d-26bddadd2fa6-hosts-file\") pod \"node-resolver-r858x\" (UID: \"de6093ed-9813-429d-812d-26bddadd2fa6\") " pod="openshift-dns/node-resolver-r858x" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774338 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774357 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774376 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774393 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774413 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774433 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774450 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774497 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774510 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772171 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772311 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772459 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772603 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772765 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772760 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.772945 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773085 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773182 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773181 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773257 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773464 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.777696 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773584 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773917 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.773955 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.774065 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.775603 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.775670 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.775764 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.775794 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.775958 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.776053 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.776072 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.776121 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.776331 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.776391 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.776546 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.777281 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.777312 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.777949 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.777379 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.777844 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.778341 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.778449 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.778753 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.778906 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.778965 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.779145 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.779346 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.779812 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.780059 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.780275 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.780307 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.780638 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.780802 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.780833 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.780695 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.781168 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.783037 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.783236 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.783257 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.783951 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.784084 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.784094 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.783724 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.784447 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.784440 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.784472 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.784455 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.784644 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.784755 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.785257 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.785294 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.785595 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.785660 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.785687 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.785863 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.786157 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.786188 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.786332 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.786779 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.786874 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.787005 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.787125 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.787304 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.787439 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.787651 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.787783 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.788008 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.788311 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.789106 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.789407 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.790124 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.791135 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.791156 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.791520 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.791864 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.793110 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.794664 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.795514 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.795768 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.796159 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.796467 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.796738 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.796776 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.797341 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.797399 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.797406 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.797423 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.797595 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.798364 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.798760 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.798769 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.799045 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.799227 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.799347 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.799490 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.800230 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.800368 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.801229 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.804111 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.805068 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.813164 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:22.313138001 +0000 UTC m=+22.219885493 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.805819 4789 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.813704 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.814387 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.814405 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.814419 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.814466 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:22.314452448 +0000 UTC m=+22.221199940 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.814567 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.814579 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.814588 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.814613 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:22.314606303 +0000 UTC m=+22.221353795 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.814647 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.815392 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.815434 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.815771 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.816113 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.816449 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.816643 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.806330 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.816754 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:22.316734753 +0000 UTC m=+22.223482245 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.809477 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.809563 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.809788 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.810037 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.810158 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.810373 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.810377 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.810392 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.810438 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.811039 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.811301 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.811339 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.811536 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.812208 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.812772 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.819476 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.819679 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.819878 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.820116 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.820233 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.820831 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.820841 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.820879 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.820906 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.821334 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.821382 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.820063 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.822242 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.825510 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.836322 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.837330 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.837961 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: E1008 14:01:21.838122 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:01:22.338101151 +0000 UTC m=+22.244848833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.838222 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.838425 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.838435 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.838550 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.838680 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.838775 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.838793 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.838919 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.839579 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.842048 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.841222 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.845528 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.846480 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.846738 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.847219 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.847554 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.847602 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.847931 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.848351 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.848422 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.848934 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.849400 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.849736 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.852382 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.853459 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.853480 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.853707 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.854060 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.854103 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.854565 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.856343 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.856525 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.857945 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.867003 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.868036 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.868301 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.868531 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.871510 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.872421 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.872596 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.872922 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.873150 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.875474 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.875798 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.875807 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-cni-binary-copy\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.875864 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-cni-binary-copy\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.875890 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.875914 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-run-k8s-cni-cncf-io\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.875932 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4w64\" (UniqueName: \"kubernetes.io/projected/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-kube-api-access-s4w64\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.875950 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2b2e313a-834b-47fb-adae-992853bf6a6a-proxy-tls\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.875968 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2b2e313a-834b-47fb-adae-992853bf6a6a-mcd-auth-proxy-config\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876012 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-cni-dir\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876038 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-socket-dir-parent\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876077 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-daemon-config\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876106 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb5jh\" (UniqueName: \"kubernetes.io/projected/2b2e313a-834b-47fb-adae-992853bf6a6a-kube-api-access-tb5jh\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876130 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-system-cni-dir\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876147 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-os-release\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876161 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-cnibin\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876174 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-etc-kubernetes\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876195 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-cnibin\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876215 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cksqm\" (UniqueName: \"kubernetes.io/projected/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-kube-api-access-cksqm\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876242 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876275 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/de6093ed-9813-429d-812d-26bddadd2fa6-hosts-file\") pod \"node-resolver-r858x\" (UID: \"de6093ed-9813-429d-812d-26bddadd2fa6\") " pod="openshift-dns/node-resolver-r858x" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876293 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-run-multus-certs\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876310 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-system-cni-dir\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876327 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-hostroot\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876352 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-run-netns\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876368 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-os-release\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876386 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-conf-dir\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876406 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876408 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2b2e313a-834b-47fb-adae-992853bf6a6a-rootfs\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876439 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876447 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/de6093ed-9813-429d-812d-26bddadd2fa6-hosts-file\") pod \"node-resolver-r858x\" (UID: \"de6093ed-9813-429d-812d-26bddadd2fa6\") " pod="openshift-dns/node-resolver-r858x" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876454 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-var-lib-cni-multus\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876468 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-var-lib-kubelet\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876523 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876539 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-var-lib-cni-bin\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876570 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5bxb\" (UniqueName: \"kubernetes.io/projected/de6093ed-9813-429d-812d-26bddadd2fa6-kube-api-access-x5bxb\") pod \"node-resolver-r858x\" (UID: \"de6093ed-9813-429d-812d-26bddadd2fa6\") " pod="openshift-dns/node-resolver-r858x" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876588 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876744 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876884 4789 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876898 4789 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876907 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876928 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876937 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876946 4789 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876954 4789 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876962 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876970 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876978 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.876999 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877008 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877016 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877025 4789 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877034 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877044 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877054 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877064 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877074 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877083 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.877092 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878069 4789 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878091 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878103 4789 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878121 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878131 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878140 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878150 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878160 4789 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878170 4789 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878181 4789 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878192 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878201 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878210 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878219 4789 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878228 4789 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878236 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878244 4789 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878252 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878260 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878268 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878276 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878285 4789 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878294 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878302 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878311 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878318 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878327 4789 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878336 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878344 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878352 4789 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878361 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878370 4789 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878380 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878391 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878401 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878411 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878420 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878429 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878438 4789 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878446 4789 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878456 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878464 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878473 4789 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878482 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878491 4789 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878499 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878508 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878516 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878525 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878534 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878543 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878553 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878561 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878570 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878579 4789 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878588 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878596 4789 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878604 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878613 4789 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878621 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878629 4789 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878637 4789 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878645 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878653 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878661 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878669 4789 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878677 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878685 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878693 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878701 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878709 4789 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878724 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878732 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878819 4789 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878832 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878842 4789 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878850 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878859 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878868 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878876 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878885 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878893 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878901 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878909 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878918 4789 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878926 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878934 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878941 4789 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878950 4789 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878959 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878967 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878976 4789 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.878997 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879005 4789 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879017 4789 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879026 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879034 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879041 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879050 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879061 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879072 4789 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879082 4789 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879091 4789 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879100 4789 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879108 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879117 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879126 4789 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879134 4789 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879143 4789 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879151 4789 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879159 4789 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879167 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879175 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879183 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879192 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879200 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879208 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879216 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879224 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879232 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879241 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879249 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879257 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879265 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879274 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879282 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879291 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879299 4789 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879307 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879315 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879323 4789 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879362 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879371 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879379 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879388 4789 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879396 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879404 4789 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879412 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879421 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879429 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879437 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879445 4789 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879452 4789 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879460 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879468 4789 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879475 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879483 4789 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879491 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879500 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879510 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879518 4789 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879527 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879534 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879542 4789 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879551 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879558 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879567 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879576 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879584 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879592 4789 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879600 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879608 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.879616 4789 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.880225 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.880648 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.883103 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.883458 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.884106 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.885061 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.910404 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5bxb\" (UniqueName: \"kubernetes.io/projected/de6093ed-9813-429d-812d-26bddadd2fa6-kube-api-access-x5bxb\") pod \"node-resolver-r858x\" (UID: \"de6093ed-9813-429d-812d-26bddadd2fa6\") " pod="openshift-dns/node-resolver-r858x" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.913300 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d"} Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.913332 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.916146 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.918567 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.919167 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.928670 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.941728 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.947123 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.967334 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980304 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-var-lib-cni-bin\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980348 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-cni-binary-copy\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980366 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-cni-binary-copy\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980384 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980400 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-run-k8s-cni-cncf-io\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980415 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4w64\" (UniqueName: \"kubernetes.io/projected/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-kube-api-access-s4w64\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980432 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2b2e313a-834b-47fb-adae-992853bf6a6a-proxy-tls\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980450 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2b2e313a-834b-47fb-adae-992853bf6a6a-mcd-auth-proxy-config\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980465 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-cni-dir\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980480 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-socket-dir-parent\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980495 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-daemon-config\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980511 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb5jh\" (UniqueName: \"kubernetes.io/projected/2b2e313a-834b-47fb-adae-992853bf6a6a-kube-api-access-tb5jh\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980527 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-system-cni-dir\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980542 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-os-release\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980558 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-cnibin\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980572 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-etc-kubernetes\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980590 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-cnibin\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980605 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cksqm\" (UniqueName: \"kubernetes.io/projected/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-kube-api-access-cksqm\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980628 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-run-multus-certs\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980643 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-system-cni-dir\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980657 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-hostroot\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980678 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-run-netns\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980694 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-os-release\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980708 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-conf-dir\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980743 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2b2e313a-834b-47fb-adae-992853bf6a6a-rootfs\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980758 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980775 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-var-lib-cni-multus\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980791 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-var-lib-kubelet\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980822 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980833 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980844 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980853 4789 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980862 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980873 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980884 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.980948 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-var-lib-kubelet\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981002 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-var-lib-cni-bin\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981050 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-cnibin\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981075 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-etc-kubernetes\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981098 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-cnibin\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981226 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-os-release\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981369 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-run-multus-certs\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981413 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-system-cni-dir\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981445 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-hostroot\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981472 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-run-netns\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981516 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-os-release\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981539 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-conf-dir\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981562 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2b2e313a-834b-47fb-adae-992853bf6a6a-rootfs\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.981830 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-cni-binary-copy\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.982099 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-var-lib-cni-multus\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.982134 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.982267 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-cni-binary-copy\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.982501 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-system-cni-dir\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.982514 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-host-run-k8s-cni-cncf-io\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.982546 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-socket-dir-parent\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.982704 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-cni-dir\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.982826 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-multus-daemon-config\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.983124 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2b2e313a-834b-47fb-adae-992853bf6a6a-mcd-auth-proxy-config\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.983170 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.989623 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.992430 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2b2e313a-834b-47fb-adae-992853bf6a6a-proxy-tls\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:21 crc kubenswrapper[4789]: I1008 14:01:21.993558 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.001212 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.004764 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cksqm\" (UniqueName: \"kubernetes.io/projected/1d006f91-6689-46a3-a6cd-7ed9f2fd9188-kube-api-access-cksqm\") pod \"multus-additional-cni-plugins-4m4jj\" (UID: \"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\") " pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.007535 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4w64\" (UniqueName: \"kubernetes.io/projected/4aee6ee4-49ed-4135-8bb0-003eb4f70f05-kube-api-access-s4w64\") pod \"multus-5428v\" (UID: \"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\") " pod="openshift-multus/multus-5428v" Oct 08 14:01:22 crc kubenswrapper[4789]: W1008 14:01:22.009551 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-37af46fdc1c7ecf35ef138ed810190a3698ea3c49575fe1725f2556f42a25494 WatchSource:0}: Error finding container 37af46fdc1c7ecf35ef138ed810190a3698ea3c49575fe1725f2556f42a25494: Status 404 returned error can't find the container with id 37af46fdc1c7ecf35ef138ed810190a3698ea3c49575fe1725f2556f42a25494 Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.013496 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.017582 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.021943 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-r858x" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.034599 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.035932 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.038349 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb5jh\" (UniqueName: \"kubernetes.io/projected/2b2e313a-834b-47fb-adae-992853bf6a6a-kube-api-access-tb5jh\") pod \"machine-config-daemon-94gx2\" (UID: \"2b2e313a-834b-47fb-adae-992853bf6a6a\") " pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:22 crc kubenswrapper[4789]: W1008 14:01:22.041508 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-d1844fccb44e799242a9eafd5cfd788fb8613f0c0d3904d569d9c1a8e12c5f1f WatchSource:0}: Error finding container d1844fccb44e799242a9eafd5cfd788fb8613f0c0d3904d569d9c1a8e12c5f1f: Status 404 returned error can't find the container with id d1844fccb44e799242a9eafd5cfd788fb8613f0c0d3904d569d9c1a8e12c5f1f Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.042264 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5428v" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.050355 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.052911 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.075327 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.090752 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9gpb2"] Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.091958 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.093163 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.096755 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.096785 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.096807 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.096934 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.097085 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.097236 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.099314 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.112497 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.130767 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.147170 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.157612 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.171782 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: W1008 14:01:22.177195 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b2e313a_834b_47fb_adae_992853bf6a6a.slice/crio-849b888004fa7edd9cebdba32d4f1eb1fb7f2385825bb7f7aa4e40f385df009b WatchSource:0}: Error finding container 849b888004fa7edd9cebdba32d4f1eb1fb7f2385825bb7f7aa4e40f385df009b: Status 404 returned error can't find the container with id 849b888004fa7edd9cebdba32d4f1eb1fb7f2385825bb7f7aa4e40f385df009b Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182224 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-netns\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182258 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-script-lib\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182280 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-config\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182298 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-slash\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182451 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-log-socket\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182565 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-openvswitch\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182598 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zkqv\" (UniqueName: \"kubernetes.io/projected/1bee958d-f383-4e01-8379-b3d2ba664129-kube-api-access-5zkqv\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182660 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-systemd\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182712 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bee958d-f383-4e01-8379-b3d2ba664129-ovn-node-metrics-cert\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182734 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-ovn-kubernetes\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182758 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-ovn\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182776 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-netd\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182861 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-node-log\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182900 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-kubelet\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182931 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-bin\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182949 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-systemd-units\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.182966 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.183044 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-var-lib-openvswitch\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.183102 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.183142 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-etc-openvswitch\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.183193 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-env-overrides\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.192867 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.206410 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.215430 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.231210 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.241821 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.257308 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.278415 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.283900 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-bin\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.283948 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-systemd-units\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.283970 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284009 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-var-lib-openvswitch\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284035 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-etc-openvswitch\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284049 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-env-overrides\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284070 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-bin\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284105 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-netns\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284124 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-script-lib\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284139 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-slash\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284149 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-var-lib-openvswitch\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284153 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-log-socket\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284172 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-config\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284188 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-openvswitch\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284204 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zkqv\" (UniqueName: \"kubernetes.io/projected/1bee958d-f383-4e01-8379-b3d2ba664129-kube-api-access-5zkqv\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284222 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-systemd\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284252 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bee958d-f383-4e01-8379-b3d2ba664129-ovn-node-metrics-cert\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284260 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-netns\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284333 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-openvswitch\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284362 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-systemd\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284336 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-ovn-kubernetes\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284447 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-log-socket\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284296 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-ovn-kubernetes\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.290554 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-etc-openvswitch\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.290598 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.284288 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-slash\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.290766 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-systemd-units\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.290940 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-script-lib\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.291207 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-ovn\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.291385 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-config\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.291684 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-ovn\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.291752 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-node-log\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.291780 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-netd\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.291805 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-kubelet\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.291885 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-kubelet\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.291963 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-node-log\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.292020 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-netd\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.294384 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bee958d-f383-4e01-8379-b3d2ba664129-ovn-node-metrics-cert\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.294416 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-env-overrides\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.313407 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zkqv\" (UniqueName: \"kubernetes.io/projected/1bee958d-f383-4e01-8379-b3d2ba664129-kube-api-access-5zkqv\") pod \"ovnkube-node-9gpb2\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.392422 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.392593 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.392701 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:01:23.392669793 +0000 UTC m=+23.299417305 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.392849 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.392759 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.392920 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.392942 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.393021 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:23.393004713 +0000 UTC m=+23.299752225 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.393068 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.393147 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:23.393133046 +0000 UTC m=+23.299880538 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.393440 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.393533 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.393581 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:23.393573129 +0000 UTC m=+23.300320621 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.393609 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.393720 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.393737 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.393746 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.393777 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:23.393770054 +0000 UTC m=+23.300517546 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.501861 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:22 crc kubenswrapper[4789]: W1008 14:01:22.513512 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bee958d_f383_4e01_8379_b3d2ba664129.slice/crio-eb0c8cc80e98c32256c7417583d52a14a2c243b48d61b6c50ca9ec08f6a05811 WatchSource:0}: Error finding container eb0c8cc80e98c32256c7417583d52a14a2c243b48d61b6c50ca9ec08f6a05811: Status 404 returned error can't find the container with id eb0c8cc80e98c32256c7417583d52a14a2c243b48d61b6c50ca9ec08f6a05811 Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.730087 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:22 crc kubenswrapper[4789]: E1008 14:01:22.730285 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.734396 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.735224 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.736400 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.737230 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.738363 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.738943 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.739597 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.740638 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.741387 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.742454 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.743095 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.744219 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.744800 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.745393 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.746335 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.746900 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.747921 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.748430 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.749063 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.750100 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.750620 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.751627 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.752141 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.753600 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.754183 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.754825 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.756084 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.756724 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.757775 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.758385 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.759387 4789 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.759590 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.761322 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.762294 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.762787 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.764323 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.765106 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.766215 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.767116 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.768204 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.768777 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.769853 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.770646 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.771695 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.772220 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.773217 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.773812 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.774976 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.775557 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.776485 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.777005 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.777970 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.778783 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.779331 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.919549 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d006f91-6689-46a3-a6cd-7ed9f2fd9188" containerID="37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4" exitCode=0 Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.919635 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" event={"ID":"1d006f91-6689-46a3-a6cd-7ed9f2fd9188","Type":"ContainerDied","Data":"37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.919715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" event={"ID":"1d006f91-6689-46a3-a6cd-7ed9f2fd9188","Type":"ContainerStarted","Data":"0ad68d5da6aee833082563a86a46d4300fab74cacfe67c7da4e2cfa558dad0cf"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.921701 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f" exitCode=0 Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.921821 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.921902 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"eb0c8cc80e98c32256c7417583d52a14a2c243b48d61b6c50ca9ec08f6a05811"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.923319 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5428v" event={"ID":"4aee6ee4-49ed-4135-8bb0-003eb4f70f05","Type":"ContainerStarted","Data":"b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.923368 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5428v" event={"ID":"4aee6ee4-49ed-4135-8bb0-003eb4f70f05","Type":"ContainerStarted","Data":"9696e61a4ed9a7c26d9fea17a4b660e253c5d498793639b3eaace1b0de3f8eed"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.925018 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-r858x" event={"ID":"de6093ed-9813-429d-812d-26bddadd2fa6","Type":"ContainerStarted","Data":"23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.925065 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-r858x" event={"ID":"de6093ed-9813-429d-812d-26bddadd2fa6","Type":"ContainerStarted","Data":"9a1e05c77a1da0f7c0925d673653b302768afdeb2f9e13e979392e90972500f7"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.928091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.928150 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.928162 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d1844fccb44e799242a9eafd5cfd788fb8613f0c0d3904d569d9c1a8e12c5f1f"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.929878 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"af5d1942e54daa2cf5163b9302fed7529d9755608a05aaf22e13f68a92c3b53b"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.933264 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.933297 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"37af46fdc1c7ecf35ef138ed810190a3698ea3c49575fe1725f2556f42a25494"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.935810 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.935902 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.935925 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"849b888004fa7edd9cebdba32d4f1eb1fb7f2385825bb7f7aa4e40f385df009b"} Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.939760 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:22Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.952686 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:22Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.966597 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:22Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:22 crc kubenswrapper[4789]: I1008 14:01:22.983646 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:22Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.004064 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.021494 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.038741 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.053428 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.068778 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.138177 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.160638 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.173477 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-tf2m8"] Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.173941 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.176266 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.176463 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.176611 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.177537 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.180668 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.193608 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.211566 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.225282 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.240395 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.255453 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.270130 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.281312 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.297111 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.302817 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zcg8\" (UniqueName: \"kubernetes.io/projected/c453426b-c6f6-4d8f-a301-5b70c2c183a1-kube-api-access-6zcg8\") pod \"node-ca-tf2m8\" (UID: \"c453426b-c6f6-4d8f-a301-5b70c2c183a1\") " pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.302864 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c453426b-c6f6-4d8f-a301-5b70c2c183a1-host\") pod \"node-ca-tf2m8\" (UID: \"c453426b-c6f6-4d8f-a301-5b70c2c183a1\") " pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.302908 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c453426b-c6f6-4d8f-a301-5b70c2c183a1-serviceca\") pod \"node-ca-tf2m8\" (UID: \"c453426b-c6f6-4d8f-a301-5b70c2c183a1\") " pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.310124 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.322682 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.338034 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.361029 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.373433 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.406569 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.406732 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.406788 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:01:25.406752586 +0000 UTC m=+25.313500078 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.406835 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c453426b-c6f6-4d8f-a301-5b70c2c183a1-serviceca\") pod \"node-ca-tf2m8\" (UID: \"c453426b-c6f6-4d8f-a301-5b70c2c183a1\") " pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.406910 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.406946 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zcg8\" (UniqueName: \"kubernetes.io/projected/c453426b-c6f6-4d8f-a301-5b70c2c183a1-kube-api-access-6zcg8\") pod \"node-ca-tf2m8\" (UID: \"c453426b-c6f6-4d8f-a301-5b70c2c183a1\") " pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.406976 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.406913 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407079 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407107 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407146 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407181 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:25.407158628 +0000 UTC m=+25.313906180 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407205 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:25.407195019 +0000 UTC m=+25.313942621 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.407019 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c453426b-c6f6-4d8f-a301-5b70c2c183a1-host\") pod \"node-ca-tf2m8\" (UID: \"c453426b-c6f6-4d8f-a301-5b70c2c183a1\") " pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407241 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407256 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407268 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.407275 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407299 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:25.407288211 +0000 UTC m=+25.314035783 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407388 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.407431 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:25.407423155 +0000 UTC m=+25.314170717 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.407651 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c453426b-c6f6-4d8f-a301-5b70c2c183a1-host\") pod \"node-ca-tf2m8\" (UID: \"c453426b-c6f6-4d8f-a301-5b70c2c183a1\") " pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.407963 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c453426b-c6f6-4d8f-a301-5b70c2c183a1-serviceca\") pod \"node-ca-tf2m8\" (UID: \"c453426b-c6f6-4d8f-a301-5b70c2c183a1\") " pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.426278 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zcg8\" (UniqueName: \"kubernetes.io/projected/c453426b-c6f6-4d8f-a301-5b70c2c183a1-kube-api-access-6zcg8\") pod \"node-ca-tf2m8\" (UID: \"c453426b-c6f6-4d8f-a301-5b70c2c183a1\") " pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.492804 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tf2m8" Oct 08 14:01:23 crc kubenswrapper[4789]: W1008 14:01:23.506401 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc453426b_c6f6_4d8f_a301_5b70c2c183a1.slice/crio-3a611402107c199291d6bcda4ad246a11c0bf96ac123b9ad65b61ded9e0b6ec7 WatchSource:0}: Error finding container 3a611402107c199291d6bcda4ad246a11c0bf96ac123b9ad65b61ded9e0b6ec7: Status 404 returned error can't find the container with id 3a611402107c199291d6bcda4ad246a11c0bf96ac123b9ad65b61ded9e0b6ec7 Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.729452 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.729541 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.729622 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:23 crc kubenswrapper[4789]: E1008 14:01:23.729766 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.940513 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" event={"ID":"1d006f91-6689-46a3-a6cd-7ed9f2fd9188","Type":"ContainerStarted","Data":"04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121"} Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.941797 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tf2m8" event={"ID":"c453426b-c6f6-4d8f-a301-5b70c2c183a1","Type":"ContainerStarted","Data":"3a611402107c199291d6bcda4ad246a11c0bf96ac123b9ad65b61ded9e0b6ec7"} Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.946264 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71"} Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.946308 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be"} Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.946318 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5"} Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.946326 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a"} Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.959889 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.973542 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:23 crc kubenswrapper[4789]: I1008 14:01:23.988481 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:23.999857 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:23Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.018609 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.032116 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.047600 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.062540 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.080612 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.099083 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.111924 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.126089 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.139457 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.732014 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:24 crc kubenswrapper[4789]: E1008 14:01:24.732511 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.950549 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379"} Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.952581 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d006f91-6689-46a3-a6cd-7ed9f2fd9188" containerID="04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121" exitCode=0 Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.952629 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" event={"ID":"1d006f91-6689-46a3-a6cd-7ed9f2fd9188","Type":"ContainerDied","Data":"04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121"} Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.954696 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tf2m8" event={"ID":"c453426b-c6f6-4d8f-a301-5b70c2c183a1","Type":"ContainerStarted","Data":"543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0"} Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.959060 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c"} Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.959111 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc"} Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.965818 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.978268 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:24 crc kubenswrapper[4789]: I1008 14:01:24.989163 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:24Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.003279 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.022092 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.033669 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.047912 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.062119 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.074919 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.093265 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.106976 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.119003 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.132698 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.144060 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.156688 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.175679 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.184855 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.196579 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.206655 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.218746 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.232283 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.244038 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.255603 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.268935 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.284229 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.298793 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.427337 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427494 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:01:29.42746817 +0000 UTC m=+29.334215662 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.427548 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.427600 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.427630 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.427667 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427736 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427759 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427774 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427811 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427837 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:29.42782029 +0000 UTC m=+29.334567782 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427854 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427818 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427915 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427927 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427898 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:29.427880422 +0000 UTC m=+29.334627914 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.427971 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:29.427963814 +0000 UTC m=+29.334711306 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.428001 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:29.427976595 +0000 UTC m=+29.334724087 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.729252 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.729791 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.729288 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:25 crc kubenswrapper[4789]: E1008 14:01:25.730197 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.964928 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d006f91-6689-46a3-a6cd-7ed9f2fd9188" containerID="12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c" exitCode=0 Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.965025 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" event={"ID":"1d006f91-6689-46a3-a6cd-7ed9f2fd9188","Type":"ContainerDied","Data":"12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c"} Oct 08 14:01:25 crc kubenswrapper[4789]: I1008 14:01:25.984242 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.001692 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:25Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.020051 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.033671 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.049259 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.060787 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.071057 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.083358 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.094403 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.110724 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.124184 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.136398 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.146316 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.359496 4789 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.361978 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.362053 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.362070 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.362214 4789 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.368111 4789 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.368314 4789 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.369121 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.369152 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.369163 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.369178 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.369189 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: E1008 14:01:26.381730 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.384912 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.384950 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.384962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.384977 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.385007 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: E1008 14:01:26.401521 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.406127 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.406188 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.406210 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.406240 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.406262 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: E1008 14:01:26.421895 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.426198 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.426231 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.426239 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.426252 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.426263 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: E1008 14:01:26.440464 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.444841 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.444887 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.444902 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.444922 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.444936 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: E1008 14:01:26.458157 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:26 crc kubenswrapper[4789]: E1008 14:01:26.458277 4789 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.460061 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.460113 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.460131 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.460153 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.460169 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.562449 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.562488 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.562500 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.562517 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.562531 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.665117 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.665161 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.665174 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.665194 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.665209 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.729549 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:26 crc kubenswrapper[4789]: E1008 14:01:26.729667 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.767393 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.767443 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.767453 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.767470 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.767502 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.871061 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.871142 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.871167 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.871200 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.871228 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.970884 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d006f91-6689-46a3-a6cd-7ed9f2fd9188" containerID="c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684" exitCode=0 Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.971023 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" event={"ID":"1d006f91-6689-46a3-a6cd-7ed9f2fd9188","Type":"ContainerDied","Data":"c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684"} Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.973619 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.973684 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.973709 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.973736 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.973764 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:26Z","lastTransitionTime":"2025-10-08T14:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.976118 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c"} Oct 08 14:01:26 crc kubenswrapper[4789]: I1008 14:01:26.996141 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:26Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.016705 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.039115 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.053582 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.076793 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.076868 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.076888 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.076917 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.076938 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:27Z","lastTransitionTime":"2025-10-08T14:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.076809 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.097317 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.114101 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.127083 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.156427 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.167844 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.180340 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.180398 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.180413 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.180437 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.180454 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:27Z","lastTransitionTime":"2025-10-08T14:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.181620 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.199701 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.210977 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.283860 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.284302 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.284321 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.284346 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.284363 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:27Z","lastTransitionTime":"2025-10-08T14:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.387648 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.387701 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.387712 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.387733 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.387746 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:27Z","lastTransitionTime":"2025-10-08T14:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.490628 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.490675 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.490686 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.490703 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.490717 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:27Z","lastTransitionTime":"2025-10-08T14:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.594387 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.594475 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.594498 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.594526 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.594547 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:27Z","lastTransitionTime":"2025-10-08T14:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.698751 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.698837 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.698852 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.698884 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.698903 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:27Z","lastTransitionTime":"2025-10-08T14:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.729265 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.729270 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:27 crc kubenswrapper[4789]: E1008 14:01:27.729503 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:27 crc kubenswrapper[4789]: E1008 14:01:27.729583 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.802273 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.802330 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.802343 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.802364 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.802379 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:27Z","lastTransitionTime":"2025-10-08T14:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.905945 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.906016 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.906035 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.906056 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.906067 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:27Z","lastTransitionTime":"2025-10-08T14:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.987256 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d006f91-6689-46a3-a6cd-7ed9f2fd9188" containerID="f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f" exitCode=0 Oct 08 14:01:27 crc kubenswrapper[4789]: I1008 14:01:27.987303 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" event={"ID":"1d006f91-6689-46a3-a6cd-7ed9f2fd9188","Type":"ContainerDied","Data":"f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.001724 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:27Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.009320 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.009379 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.009406 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.009441 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.009470 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.020128 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.042401 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.057674 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.078933 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.098799 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.115025 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.115197 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.115209 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.115227 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.115239 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.117091 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.132316 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.150303 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.163185 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.173409 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.186468 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.199012 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.217848 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.217904 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.217917 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.217938 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.217953 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.320909 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.320964 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.320975 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.321022 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.321036 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.424097 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.424173 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.424185 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.424207 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.424218 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.527223 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.527283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.527304 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.527331 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.527350 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.629574 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.629608 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.629616 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.629630 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.629641 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.729789 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:28 crc kubenswrapper[4789]: E1008 14:01:28.730063 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.737360 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.737390 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.737400 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.737416 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.737425 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.839678 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.839729 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.839741 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.839764 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.839778 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.943385 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.943438 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.943457 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.943477 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.943493 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:28Z","lastTransitionTime":"2025-10-08T14:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.998552 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317"} Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.998954 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:28 crc kubenswrapper[4789]: I1008 14:01:28.999008 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.002314 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d006f91-6689-46a3-a6cd-7ed9f2fd9188" containerID="50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431" exitCode=0 Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.002356 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" event={"ID":"1d006f91-6689-46a3-a6cd-7ed9f2fd9188","Type":"ContainerDied","Data":"50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.015242 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.029798 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.031498 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.046601 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.046634 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.046643 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.046657 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.046667 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.049454 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.064929 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.084661 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.103838 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.121078 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.135850 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.149654 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.149693 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.149704 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.149725 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.149738 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.155482 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.171805 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.188505 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.202922 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.220584 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.236470 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.253166 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.253242 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.253261 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.253291 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.253310 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.255074 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.267745 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.281629 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.295102 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.307909 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.320916 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.330809 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.345774 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.355976 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.356048 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.356062 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.356080 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.356091 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.356675 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.368643 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.391142 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.403070 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:29Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.458760 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.458808 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.458820 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.458837 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.458850 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.468499 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.468632 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468648 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:01:37.468627969 +0000 UTC m=+37.375375461 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.468704 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.468748 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.468790 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468755 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468816 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468842 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468859 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468868 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:37.468853986 +0000 UTC m=+37.375601478 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468808 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468899 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:37.468886207 +0000 UTC m=+37.375633689 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468930 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:37.468919098 +0000 UTC m=+37.375666700 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.468968 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.469098 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.469122 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.469215 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:37.469193715 +0000 UTC m=+37.375941247 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.562129 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.562206 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.562225 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.562250 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.562267 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.664916 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.665026 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.665047 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.665076 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.665095 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.729266 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.729266 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.729505 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:29 crc kubenswrapper[4789]: E1008 14:01:29.729601 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.767905 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.767948 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.767959 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.767976 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.768007 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.870874 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.870935 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.870951 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.870969 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.871006 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.975391 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.975458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.975475 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.975504 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:29 crc kubenswrapper[4789]: I1008 14:01:29.975525 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:29Z","lastTransitionTime":"2025-10-08T14:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.013451 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" event={"ID":"1d006f91-6689-46a3-a6cd-7ed9f2fd9188","Type":"ContainerStarted","Data":"bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.014170 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.036205 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.053219 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.057052 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.076330 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.077904 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.077973 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.078048 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.078078 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.078097 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:30Z","lastTransitionTime":"2025-10-08T14:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.096875 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.113692 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.134448 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.151242 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.168388 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.180508 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.180677 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.180703 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.180738 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.180760 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:30Z","lastTransitionTime":"2025-10-08T14:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.184029 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.196786 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.211615 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.227542 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.251924 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.266317 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.284073 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.284127 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.284145 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.284173 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.284194 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:30Z","lastTransitionTime":"2025-10-08T14:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.288390 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.305779 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.321696 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.339193 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.356024 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.369231 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.382703 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.387123 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.387209 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.387234 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.387266 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.387287 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:30Z","lastTransitionTime":"2025-10-08T14:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.395403 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.407953 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.431427 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.452350 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.468859 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.490402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.490465 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.490483 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.490515 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.490533 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:30Z","lastTransitionTime":"2025-10-08T14:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.593798 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.593869 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.593881 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.593938 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.593955 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:30Z","lastTransitionTime":"2025-10-08T14:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.696959 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.697058 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.697079 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.697108 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.697127 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:30Z","lastTransitionTime":"2025-10-08T14:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.729685 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:30 crc kubenswrapper[4789]: E1008 14:01:30.729890 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.746838 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.762199 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.777445 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.792063 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.800654 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.800708 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.800719 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.800739 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.800753 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:30Z","lastTransitionTime":"2025-10-08T14:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.808949 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.826398 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.847265 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.863845 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.880976 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.900856 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.903822 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.903863 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.903872 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.903892 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.903904 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:30Z","lastTransitionTime":"2025-10-08T14:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.919505 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.952161 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:30 crc kubenswrapper[4789]: I1008 14:01:30.968649 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.008569 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.008606 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.008616 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.008633 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.008644 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.117295 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.117321 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.117328 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.117342 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.117352 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.220142 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.220190 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.220205 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.220227 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.220237 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.315721 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.322585 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.322624 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.322633 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.322651 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.322662 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.336004 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.355250 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.375755 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.387380 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.409659 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.425041 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.425091 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.425100 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.425123 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.425135 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.430784 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.445438 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.459552 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.474839 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.506762 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.528121 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.528169 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.528196 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.528243 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.528290 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.530505 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.548849 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.566044 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.633336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.633396 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.633415 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.633442 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.633466 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.730105 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.730111 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:31 crc kubenswrapper[4789]: E1008 14:01:31.730952 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:31 crc kubenswrapper[4789]: E1008 14:01:31.731148 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.736097 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.736149 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.736164 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.736187 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.736203 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.838778 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.838818 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.838828 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.838845 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.838860 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.942052 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.942128 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.942142 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.942162 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:31 crc kubenswrapper[4789]: I1008 14:01:31.942206 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:31Z","lastTransitionTime":"2025-10-08T14:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.026286 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/0.log" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.030522 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317" exitCode=1 Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.030579 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.031905 4789 scope.go:117] "RemoveContainer" containerID="5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.045310 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.045378 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.045397 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.045426 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.045448 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.056630 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.079763 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.096030 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.115218 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.132374 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.148113 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.149577 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.149658 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.149680 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.149709 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.149729 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.163508 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.179120 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.193016 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.210552 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.229846 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.253374 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.253421 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.253434 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.253458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.253471 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.253622 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"message\\\":\\\"/factory.go:140\\\\nI1008 14:01:31.818844 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:01:31.818861 6111 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819367 6111 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819627 6111 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819723 6111 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 14:01:31.819881 6111 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.820311 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 14:01:31.820357 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:01:31.820373 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 14:01:31.820404 6111 factory.go:656] Stopping watch factory\\\\nI1008 14:01:31.820429 6111 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.268294 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:32Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.356442 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.356490 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.356501 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.356524 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.356536 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.460189 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.460271 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.460294 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.460322 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.460341 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.562784 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.562875 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.562893 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.562928 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.562948 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.664797 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.664838 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.664850 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.664867 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.664878 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.729877 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:32 crc kubenswrapper[4789]: E1008 14:01:32.730170 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.768217 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.768305 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.768325 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.768353 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.768372 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.872145 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.872219 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.872240 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.872270 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.872292 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.975759 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.975847 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.975866 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.975895 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:32 crc kubenswrapper[4789]: I1008 14:01:32.975914 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:32Z","lastTransitionTime":"2025-10-08T14:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.038270 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/0.log" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.044273 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.045059 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.065631 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.078838 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.078886 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.078897 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.078916 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.078928 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:33Z","lastTransitionTime":"2025-10-08T14:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.086069 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.103551 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.127392 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.149546 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.182469 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.182532 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.182551 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.182576 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.182593 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:33Z","lastTransitionTime":"2025-10-08T14:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.188693 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.212425 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.238942 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.253338 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.265213 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.276397 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.284763 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.284810 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.284826 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.284848 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.284860 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:33Z","lastTransitionTime":"2025-10-08T14:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.295440 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"message\\\":\\\"/factory.go:140\\\\nI1008 14:01:31.818844 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:01:31.818861 6111 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819367 6111 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819627 6111 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819723 6111 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 14:01:31.819881 6111 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.820311 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 14:01:31.820357 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:01:31.820373 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 14:01:31.820404 6111 factory.go:656] Stopping watch factory\\\\nI1008 14:01:31.820429 6111 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.305687 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:33Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.387705 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.387737 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.387748 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.387773 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.387784 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:33Z","lastTransitionTime":"2025-10-08T14:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.490594 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.490656 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.490670 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.490692 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.490705 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:33Z","lastTransitionTime":"2025-10-08T14:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.594141 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.594273 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.594295 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.594332 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.594352 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:33Z","lastTransitionTime":"2025-10-08T14:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.697354 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.697473 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.697488 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.697511 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.697530 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:33Z","lastTransitionTime":"2025-10-08T14:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.729276 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:33 crc kubenswrapper[4789]: E1008 14:01:33.729421 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.729276 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:33 crc kubenswrapper[4789]: E1008 14:01:33.729521 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.801737 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.801785 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.801796 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.801818 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.801830 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:33Z","lastTransitionTime":"2025-10-08T14:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.905290 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.905361 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.905378 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.905410 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:33 crc kubenswrapper[4789]: I1008 14:01:33.905429 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:33Z","lastTransitionTime":"2025-10-08T14:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.008276 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.008321 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.008333 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.008351 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.008363 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.052771 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/1.log" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.054081 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/0.log" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.058585 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c" exitCode=1 Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.058658 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.058737 4789 scope.go:117] "RemoveContainer" containerID="5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.059908 4789 scope.go:117] "RemoveContainer" containerID="5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c" Oct 08 14:01:34 crc kubenswrapper[4789]: E1008 14:01:34.060207 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.082663 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.097623 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr"] Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.098700 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.100925 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.101363 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.102617 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.111421 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.111483 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.111494 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.111527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.111538 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.122455 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.134270 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.156588 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.171937 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.192079 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.207835 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.213291 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.213332 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.213344 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.213359 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.213369 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.224309 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.224483 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.224569 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqgdp\" (UniqueName: \"kubernetes.io/projected/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-kube-api-access-vqgdp\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.224717 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.229304 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"message\\\":\\\"/factory.go:140\\\\nI1008 14:01:31.818844 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:01:31.818861 6111 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819367 6111 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819627 6111 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819723 6111 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 14:01:31.819881 6111 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.820311 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 14:01:31.820357 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:01:31.820373 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 14:01:31.820404 6111 factory.go:656] Stopping watch factory\\\\nI1008 14:01:31.820429 6111 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:33Z\\\",\\\"message\\\":\\\"dler 7\\\\nI1008 14:01:32.909667 6235 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909660 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1008 14:01:32.909711 6235 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 2.046028ms\\\\nI1008 14:01:32.909737 6235 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1008 14:01:32.909737 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 14:01:32.909784 6235 factory.go:656] Stopping watch factory\\\\nI1008 14:01:32.909805 6235 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:01:32.909785 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nI1008 14:01:32.909874 6235 services_controller.go:360] Finished syncing service machine-api-operator-webhook on namespace openshift-machine-api for network=default : 2.387968ms\\\\nI1008 14:01:32.909851 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909853 6235 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:01:32.910156 6235 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.242547 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.255795 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.271155 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.282831 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.300542 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.316061 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.316103 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.316116 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.316135 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.316149 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.325621 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.325707 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.325757 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqgdp\" (UniqueName: \"kubernetes.io/projected/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-kube-api-access-vqgdp\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.325808 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.326618 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.326645 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.327531 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.337374 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.342493 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.348740 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqgdp\" (UniqueName: \"kubernetes.io/projected/95e63cbd-0d26-4d8e-b6e2-768355e7dcca-kube-api-access-vqgdp\") pod \"ovnkube-control-plane-749d76644c-xhhmr\" (UID: \"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.357543 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.370172 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.382704 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.394890 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.406384 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.418450 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.418510 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.418554 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.418568 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.418591 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.418607 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.420340 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: W1008 14:01:34.432861 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95e63cbd_0d26_4d8e_b6e2_768355e7dcca.slice/crio-b43be5d7f6e464dc0d7f7090e75461034b7d6061172720c1f279e2509a921e00 WatchSource:0}: Error finding container b43be5d7f6e464dc0d7f7090e75461034b7d6061172720c1f279e2509a921e00: Status 404 returned error can't find the container with id b43be5d7f6e464dc0d7f7090e75461034b7d6061172720c1f279e2509a921e00 Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.437093 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.461105 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"message\\\":\\\"/factory.go:140\\\\nI1008 14:01:31.818844 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:01:31.818861 6111 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819367 6111 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819627 6111 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819723 6111 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 14:01:31.819881 6111 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.820311 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 14:01:31.820357 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:01:31.820373 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 14:01:31.820404 6111 factory.go:656] Stopping watch factory\\\\nI1008 14:01:31.820429 6111 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:33Z\\\",\\\"message\\\":\\\"dler 7\\\\nI1008 14:01:32.909667 6235 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909660 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1008 14:01:32.909711 6235 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 2.046028ms\\\\nI1008 14:01:32.909737 6235 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1008 14:01:32.909737 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 14:01:32.909784 6235 factory.go:656] Stopping watch factory\\\\nI1008 14:01:32.909805 6235 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:01:32.909785 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nI1008 14:01:32.909874 6235 services_controller.go:360] Finished syncing service machine-api-operator-webhook on namespace openshift-machine-api for network=default : 2.387968ms\\\\nI1008 14:01:32.909851 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909853 6235 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:01:32.910156 6235 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.475472 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.494943 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.506405 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.520818 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.521101 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.521118 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.521132 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.521141 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.623608 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.623647 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.623662 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.623679 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.623689 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.726788 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.726845 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.726858 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.726903 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.726916 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.729401 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:34 crc kubenswrapper[4789]: E1008 14:01:34.729533 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.818860 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-b269s"] Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.819505 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:34 crc kubenswrapper[4789]: E1008 14:01:34.819593 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.829882 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.829918 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.829929 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.829948 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.829963 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.839611 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.853408 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.865933 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.877359 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.890703 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.908779 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.921604 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.931266 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.931314 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t54j9\" (UniqueName: \"kubernetes.io/projected/765705a4-a303-4281-9677-5d0769f6d157-kube-api-access-t54j9\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.932516 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.932543 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.932557 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.932576 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.932589 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:34Z","lastTransitionTime":"2025-10-08T14:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.934412 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.955152 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"message\\\":\\\"/factory.go:140\\\\nI1008 14:01:31.818844 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:01:31.818861 6111 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819367 6111 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819627 6111 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819723 6111 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 14:01:31.819881 6111 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.820311 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 14:01:31.820357 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:01:31.820373 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 14:01:31.820404 6111 factory.go:656] Stopping watch factory\\\\nI1008 14:01:31.820429 6111 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:33Z\\\",\\\"message\\\":\\\"dler 7\\\\nI1008 14:01:32.909667 6235 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909660 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1008 14:01:32.909711 6235 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 2.046028ms\\\\nI1008 14:01:32.909737 6235 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1008 14:01:32.909737 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 14:01:32.909784 6235 factory.go:656] Stopping watch factory\\\\nI1008 14:01:32.909805 6235 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:01:32.909785 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nI1008 14:01:32.909874 6235 services_controller.go:360] Finished syncing service machine-api-operator-webhook on namespace openshift-machine-api for network=default : 2.387968ms\\\\nI1008 14:01:32.909851 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909853 6235 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:01:32.910156 6235 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.969404 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:34 crc kubenswrapper[4789]: I1008 14:01:34.986243 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:34Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.002387 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.019204 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.031773 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.031816 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t54j9\" (UniqueName: \"kubernetes.io/projected/765705a4-a303-4281-9677-5d0769f6d157-kube-api-access-t54j9\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:35 crc kubenswrapper[4789]: E1008 14:01:35.031933 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:35 crc kubenswrapper[4789]: E1008 14:01:35.032005 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs podName:765705a4-a303-4281-9677-5d0769f6d157 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:35.531973514 +0000 UTC m=+35.438721006 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs") pod "network-metrics-daemon-b269s" (UID: "765705a4-a303-4281-9677-5d0769f6d157") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.032643 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.034397 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.034427 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.034449 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.034463 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.034472 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.043686 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.052169 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t54j9\" (UniqueName: \"kubernetes.io/projected/765705a4-a303-4281-9677-5d0769f6d157-kube-api-access-t54j9\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.063849 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" event={"ID":"95e63cbd-0d26-4d8e-b6e2-768355e7dcca","Type":"ContainerStarted","Data":"72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.063896 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" event={"ID":"95e63cbd-0d26-4d8e-b6e2-768355e7dcca","Type":"ContainerStarted","Data":"2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.063908 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" event={"ID":"95e63cbd-0d26-4d8e-b6e2-768355e7dcca","Type":"ContainerStarted","Data":"b43be5d7f6e464dc0d7f7090e75461034b7d6061172720c1f279e2509a921e00"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.065876 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/1.log" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.069465 4789 scope.go:117] "RemoveContainer" containerID="5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c" Oct 08 14:01:35 crc kubenswrapper[4789]: E1008 14:01:35.069768 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.079314 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.090580 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.106326 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.125586 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.136523 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.136568 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.136580 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.136598 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.136611 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.143769 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.158006 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.173264 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.190821 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e9ef45c7a2124da977d70dc3f5e95a16eea39fc32e4275cd5953232ff924317\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"message\\\":\\\"/factory.go:140\\\\nI1008 14:01:31.818844 6111 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:01:31.818861 6111 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819367 6111 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819627 6111 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.819723 6111 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 14:01:31.819881 6111 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:01:31.820311 6111 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 14:01:31.820357 6111 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:01:31.820373 6111 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 14:01:31.820404 6111 factory.go:656] Stopping watch factory\\\\nI1008 14:01:31.820429 6111 ovnkube.go:599] Stopped ovnkube\\\\nI1008 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:33Z\\\",\\\"message\\\":\\\"dler 7\\\\nI1008 14:01:32.909667 6235 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909660 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1008 14:01:32.909711 6235 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 2.046028ms\\\\nI1008 14:01:32.909737 6235 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1008 14:01:32.909737 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 14:01:32.909784 6235 factory.go:656] Stopping watch factory\\\\nI1008 14:01:32.909805 6235 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:01:32.909785 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nI1008 14:01:32.909874 6235 services_controller.go:360] Finished syncing service machine-api-operator-webhook on namespace openshift-machine-api for network=default : 2.387968ms\\\\nI1008 14:01:32.909851 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909853 6235 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:01:32.910156 6235 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.201663 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.213117 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.227812 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.238850 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.238924 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.238954 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.238981 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.239032 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.248491 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.263273 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.274352 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.288033 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.300854 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.316078 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.327382 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.341515 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.341559 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.341568 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.341585 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.341596 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.342731 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.356278 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.369742 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.380848 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.393043 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.412271 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:33Z\\\",\\\"message\\\":\\\"dler 7\\\\nI1008 14:01:32.909667 6235 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909660 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1008 14:01:32.909711 6235 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 2.046028ms\\\\nI1008 14:01:32.909737 6235 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1008 14:01:32.909737 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 14:01:32.909784 6235 factory.go:656] Stopping watch factory\\\\nI1008 14:01:32.909805 6235 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:01:32.909785 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nI1008 14:01:32.909874 6235 services_controller.go:360] Finished syncing service machine-api-operator-webhook on namespace openshift-machine-api for network=default : 2.387968ms\\\\nI1008 14:01:32.909851 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909853 6235 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:01:32.910156 6235 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.422066 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.439146 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.444170 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.444254 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.444270 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.444290 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.444302 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.453364 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.468497 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.482497 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.493849 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:35Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.537067 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:35 crc kubenswrapper[4789]: E1008 14:01:35.537340 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:35 crc kubenswrapper[4789]: E1008 14:01:35.537479 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs podName:765705a4-a303-4281-9677-5d0769f6d157 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:36.537445851 +0000 UTC m=+36.444193383 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs") pod "network-metrics-daemon-b269s" (UID: "765705a4-a303-4281-9677-5d0769f6d157") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.548061 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.548134 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.548162 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.548197 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.548224 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.651921 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.651975 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.652029 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.652060 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.652082 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.729509 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.729631 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:35 crc kubenswrapper[4789]: E1008 14:01:35.729735 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:35 crc kubenswrapper[4789]: E1008 14:01:35.729898 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.755344 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.755380 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.755389 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.755403 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.755414 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.858312 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.858378 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.858396 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.858422 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.858440 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.962457 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.962546 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.962566 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.962593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:35 crc kubenswrapper[4789]: I1008 14:01:35.962613 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:35Z","lastTransitionTime":"2025-10-08T14:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.066325 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.066388 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.066411 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.066444 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.066466 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.169908 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.169955 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.169964 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.169983 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.170015 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.273354 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.273424 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.273443 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.273467 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.273509 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.376358 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.376417 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.376434 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.376459 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.376479 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.478948 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.478975 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.478982 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.479013 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.479024 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.548047 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.548195 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.548249 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs podName:765705a4-a303-4281-9677-5d0769f6d157 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:38.548235739 +0000 UTC m=+38.454983231 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs") pod "network-metrics-daemon-b269s" (UID: "765705a4-a303-4281-9677-5d0769f6d157") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.574517 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.574589 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.574614 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.574643 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.574666 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.589740 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:36Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.594016 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.594075 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.594094 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.594115 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.594132 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.610797 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:36Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.615346 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.615407 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.615424 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.615447 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.615463 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.633783 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:36Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.638828 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.638896 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.638921 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.638954 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.638976 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.658443 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:36Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.662964 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.663017 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.663028 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.663046 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.663058 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.676136 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:36Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.676356 4789 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.678124 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.678175 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.678192 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.678215 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.678233 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.729473 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.729642 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.729869 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:36 crc kubenswrapper[4789]: E1008 14:01:36.730076 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.781438 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.781515 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.781533 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.781562 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.781586 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.885014 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.885089 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.885107 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.885139 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.885158 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.988586 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.988646 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.988696 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.988740 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:36 crc kubenswrapper[4789]: I1008 14:01:36.988758 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:36Z","lastTransitionTime":"2025-10-08T14:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.091136 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.091193 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.091210 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.091234 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.091252 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:37Z","lastTransitionTime":"2025-10-08T14:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.193966 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.194070 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.194088 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.194113 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.194131 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:37Z","lastTransitionTime":"2025-10-08T14:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.298169 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.298338 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.298358 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.298386 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.298404 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:37Z","lastTransitionTime":"2025-10-08T14:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.402332 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.402393 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.402410 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.402440 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.402457 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:37Z","lastTransitionTime":"2025-10-08T14:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.505658 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.505701 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.505710 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.505725 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.505738 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:37Z","lastTransitionTime":"2025-10-08T14:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.561765 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562030 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:01:53.561902951 +0000 UTC m=+53.468650493 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.562100 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.562175 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562292 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.562280 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562395 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:53.562365704 +0000 UTC m=+53.469113236 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.562463 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562517 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562549 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562575 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562580 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562642 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562657 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:53.562634761 +0000 UTC m=+53.469382293 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562668 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562762 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:53.562730094 +0000 UTC m=+53.469477636 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562601 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.562866 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:53.562847868 +0000 UTC m=+53.469595390 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.608526 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.608587 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.608604 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.608630 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.608649 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:37Z","lastTransitionTime":"2025-10-08T14:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.712219 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.712278 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.712297 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.712322 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.712381 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:37Z","lastTransitionTime":"2025-10-08T14:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.729652 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.729745 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.729831 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:37 crc kubenswrapper[4789]: E1008 14:01:37.729925 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.815901 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.815975 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.816041 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.816074 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.816101 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:37Z","lastTransitionTime":"2025-10-08T14:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.919193 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.919263 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.919280 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.919306 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:37 crc kubenswrapper[4789]: I1008 14:01:37.919323 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:37Z","lastTransitionTime":"2025-10-08T14:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.022361 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.022428 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.022446 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.022476 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.022493 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.125775 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.125833 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.125849 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.125873 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.125891 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.229415 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.229774 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.229792 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.229823 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.229840 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.331917 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.331964 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.331976 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.332016 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.332034 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.435056 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.435093 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.435104 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.435120 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.435132 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.537509 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.537547 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.537558 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.537574 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.537585 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.573146 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:38 crc kubenswrapper[4789]: E1008 14:01:38.573317 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:38 crc kubenswrapper[4789]: E1008 14:01:38.573419 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs podName:765705a4-a303-4281-9677-5d0769f6d157 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:42.573394031 +0000 UTC m=+42.480141553 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs") pod "network-metrics-daemon-b269s" (UID: "765705a4-a303-4281-9677-5d0769f6d157") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.640790 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.640850 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.640867 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.640893 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.640923 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.729209 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.729265 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:38 crc kubenswrapper[4789]: E1008 14:01:38.729352 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:38 crc kubenswrapper[4789]: E1008 14:01:38.729525 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.744554 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.744594 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.744603 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.744633 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.744645 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.847073 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.847147 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.847168 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.847198 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.847217 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.950327 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.950377 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.950386 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.950402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:38 crc kubenswrapper[4789]: I1008 14:01:38.950412 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:38Z","lastTransitionTime":"2025-10-08T14:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.053373 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.053425 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.053436 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.053456 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.053469 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.163383 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.163576 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.163622 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.163665 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.163785 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.267368 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.267512 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.267556 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.267585 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.267604 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.371727 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.371881 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.371909 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.371942 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.371963 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.475254 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.475300 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.475311 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.475330 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.475342 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.578171 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.578210 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.578218 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.578232 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.578240 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.681587 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.681651 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.681668 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.681693 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.681713 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.729569 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.729623 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:39 crc kubenswrapper[4789]: E1008 14:01:39.729779 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:39 crc kubenswrapper[4789]: E1008 14:01:39.729885 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.785490 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.785550 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.785573 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.785600 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.785618 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.888626 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.888675 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.888690 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.888716 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.888732 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.991687 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.991747 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.991763 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.991785 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:39 crc kubenswrapper[4789]: I1008 14:01:39.991799 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:39Z","lastTransitionTime":"2025-10-08T14:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.093580 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.093657 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.093677 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.093706 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.093729 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:40Z","lastTransitionTime":"2025-10-08T14:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.197169 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.197233 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.197251 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.197282 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.197301 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:40Z","lastTransitionTime":"2025-10-08T14:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.300527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.300640 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.300665 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.300739 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.300767 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:40Z","lastTransitionTime":"2025-10-08T14:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.403844 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.403913 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.403925 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.403948 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.403978 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:40Z","lastTransitionTime":"2025-10-08T14:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.506703 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.506765 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.506773 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.506787 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.506797 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:40Z","lastTransitionTime":"2025-10-08T14:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.609149 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.609229 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.609251 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.609281 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.609302 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:40Z","lastTransitionTime":"2025-10-08T14:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.713099 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.713133 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.713141 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.713157 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.713167 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:40Z","lastTransitionTime":"2025-10-08T14:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.729595 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:40 crc kubenswrapper[4789]: E1008 14:01:40.729705 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.729744 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:40 crc kubenswrapper[4789]: E1008 14:01:40.729883 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.744211 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.765208 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.784043 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.801401 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.816478 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.816519 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.816533 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.816555 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.816572 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:40Z","lastTransitionTime":"2025-10-08T14:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.819194 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.834980 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.857700 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.870967 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.889730 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.919093 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.919144 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.919162 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.919185 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.919204 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:40Z","lastTransitionTime":"2025-10-08T14:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.924936 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:33Z\\\",\\\"message\\\":\\\"dler 7\\\\nI1008 14:01:32.909667 6235 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909660 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1008 14:01:32.909711 6235 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 2.046028ms\\\\nI1008 14:01:32.909737 6235 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1008 14:01:32.909737 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 14:01:32.909784 6235 factory.go:656] Stopping watch factory\\\\nI1008 14:01:32.909805 6235 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:01:32.909785 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nI1008 14:01:32.909874 6235 services_controller.go:360] Finished syncing service machine-api-operator-webhook on namespace openshift-machine-api for network=default : 2.387968ms\\\\nI1008 14:01:32.909851 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909853 6235 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:01:32.910156 6235 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.942739 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.959083 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.972202 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.985473 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:40 crc kubenswrapper[4789]: I1008 14:01:40.997493 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:40Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.021478 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.021541 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.021560 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.021586 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.021607 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.123763 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.123841 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.123872 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.123905 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.123929 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.227366 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.227411 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.227426 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.227446 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.227461 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.330293 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.330999 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.331129 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.331217 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.331324 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.434354 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.434402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.434412 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.434428 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.434439 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.537050 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.537080 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.537088 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.537101 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.537110 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.639195 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.639228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.639238 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.639253 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.639262 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.729433 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.729433 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:41 crc kubenswrapper[4789]: E1008 14:01:41.729582 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:41 crc kubenswrapper[4789]: E1008 14:01:41.729659 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.741437 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.741573 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.741589 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.741603 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.741611 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.843731 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.843765 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.843775 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.843791 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.843802 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.946343 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.946401 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.946416 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.946432 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:41 crc kubenswrapper[4789]: I1008 14:01:41.946442 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:41Z","lastTransitionTime":"2025-10-08T14:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.049091 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.049162 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.049184 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.049211 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.049229 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.152689 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.152781 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.152815 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.152849 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.152871 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.262368 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.262444 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.262467 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.262497 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.262517 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.365259 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.365319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.365333 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.365356 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.365371 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.468848 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.468927 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.468950 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.468980 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.469045 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.571480 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.571773 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.571916 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.572054 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.572162 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.614790 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:42 crc kubenswrapper[4789]: E1008 14:01:42.615023 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:42 crc kubenswrapper[4789]: E1008 14:01:42.615341 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs podName:765705a4-a303-4281-9677-5d0769f6d157 nodeName:}" failed. No retries permitted until 2025-10-08 14:01:50.615307773 +0000 UTC m=+50.522055305 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs") pod "network-metrics-daemon-b269s" (UID: "765705a4-a303-4281-9677-5d0769f6d157") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.674787 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.674834 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.674845 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.674862 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.674897 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.730036 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.730050 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:42 crc kubenswrapper[4789]: E1008 14:01:42.730434 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:42 crc kubenswrapper[4789]: E1008 14:01:42.730687 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.777462 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.777514 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.777526 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.777544 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.777556 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.880022 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.880077 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.880088 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.880110 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.880123 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.982703 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.982739 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.982765 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.982783 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:42 crc kubenswrapper[4789]: I1008 14:01:42.982793 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:42Z","lastTransitionTime":"2025-10-08T14:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.085106 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.085148 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.085157 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.085174 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.085183 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:43Z","lastTransitionTime":"2025-10-08T14:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.187904 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.187960 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.187969 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.187989 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.188024 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:43Z","lastTransitionTime":"2025-10-08T14:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.289947 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.290012 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.290024 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.290042 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.290057 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:43Z","lastTransitionTime":"2025-10-08T14:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.392731 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.392773 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.392781 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.392796 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.392805 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:43Z","lastTransitionTime":"2025-10-08T14:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.495548 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.495616 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.495630 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.495667 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.495683 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:43Z","lastTransitionTime":"2025-10-08T14:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.598149 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.598219 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.598236 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.598267 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.598292 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:43Z","lastTransitionTime":"2025-10-08T14:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.700952 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.701032 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.701045 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.701063 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.701076 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:43Z","lastTransitionTime":"2025-10-08T14:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.729241 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:43 crc kubenswrapper[4789]: E1008 14:01:43.729417 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.729263 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:43 crc kubenswrapper[4789]: E1008 14:01:43.729541 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.804462 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.804544 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.804581 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.804626 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.804649 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:43Z","lastTransitionTime":"2025-10-08T14:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.908208 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.908265 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.908281 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.908305 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:43 crc kubenswrapper[4789]: I1008 14:01:43.908323 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:43Z","lastTransitionTime":"2025-10-08T14:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.011831 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.011878 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.011896 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.011915 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.011928 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.115074 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.115116 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.115125 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.115145 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.115157 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.217579 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.217623 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.217634 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.217652 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.217665 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.320200 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.320263 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.320279 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.320305 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.320322 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.423200 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.423283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.423300 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.423326 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.423347 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.526388 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.526458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.526482 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.526515 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.526539 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.630183 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.630249 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.630260 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.630278 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.630292 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.730085 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.730196 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:44 crc kubenswrapper[4789]: E1008 14:01:44.730298 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:44 crc kubenswrapper[4789]: E1008 14:01:44.730396 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.733374 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.733463 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.733484 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.733509 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.733527 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.836512 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.836582 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.836603 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.836631 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.836654 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.939202 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.939267 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.939282 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.939308 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:44 crc kubenswrapper[4789]: I1008 14:01:44.939330 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:44Z","lastTransitionTime":"2025-10-08T14:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.042603 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.042645 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.042658 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.042676 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.042687 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.145500 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.145563 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.145580 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.145609 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.145627 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.248972 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.249072 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.249089 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.249115 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.249136 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.352217 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.352298 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.352322 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.352356 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.352383 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.455656 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.455837 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.455862 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.455888 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.455908 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.560305 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.560354 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.560365 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.560383 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.560397 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.663336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.663409 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.663429 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.663459 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.663480 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.729494 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.729550 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:45 crc kubenswrapper[4789]: E1008 14:01:45.729689 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:45 crc kubenswrapper[4789]: E1008 14:01:45.729851 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.766904 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.766962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.766981 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.767036 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.767054 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.869582 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.869624 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.869634 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.869651 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.869663 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.972754 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.972820 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.972832 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.972849 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:45 crc kubenswrapper[4789]: I1008 14:01:45.972861 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:45Z","lastTransitionTime":"2025-10-08T14:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.075200 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.075275 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.075294 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.075322 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.075341 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.178613 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.178682 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.178705 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.178735 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.178756 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.281852 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.281913 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.281937 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.281967 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.282025 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.385167 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.385235 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.385256 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.385285 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.385312 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.488790 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.488849 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.488866 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.488892 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.488909 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.591338 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.591383 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.591394 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.591411 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.591422 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.694028 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.694072 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.694083 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.694101 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.694112 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.729975 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.730181 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:46 crc kubenswrapper[4789]: E1008 14:01:46.730443 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:46 crc kubenswrapper[4789]: E1008 14:01:46.730562 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.731674 4789 scope.go:117] "RemoveContainer" containerID="5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.797103 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.797137 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.797146 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.797162 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.797172 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.899737 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.899779 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.899791 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.899809 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.899822 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.961871 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.961921 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.961937 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.961960 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.961978 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: E1008 14:01:46.978141 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:46Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.981597 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.981650 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.981660 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.981677 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.981687 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:46 crc kubenswrapper[4789]: E1008 14:01:46.993354 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:46Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.997229 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.997279 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.997295 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.997319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:46 crc kubenswrapper[4789]: I1008 14:01:46.997336 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:46Z","lastTransitionTime":"2025-10-08T14:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: E1008 14:01:47.011455 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.015414 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.015491 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.015514 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.015592 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.015636 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: E1008 14:01:47.030038 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.034479 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.034533 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.034547 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.034566 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.034578 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: E1008 14:01:47.048752 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: E1008 14:01:47.048899 4789 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.050741 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.050776 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.050785 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.050801 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.050810 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.116347 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/1.log" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.119256 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.119688 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.132698 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.152860 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.153166 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.153216 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.153228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.153246 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.153258 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.166999 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.180792 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.191276 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.204825 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.219704 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.232323 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.246589 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.255197 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.255234 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.255243 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.255263 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.255274 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.264080 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:33Z\\\",\\\"message\\\":\\\"dler 7\\\\nI1008 14:01:32.909667 6235 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909660 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1008 14:01:32.909711 6235 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 2.046028ms\\\\nI1008 14:01:32.909737 6235 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1008 14:01:32.909737 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 14:01:32.909784 6235 factory.go:656] Stopping watch factory\\\\nI1008 14:01:32.909805 6235 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:01:32.909785 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nI1008 14:01:32.909874 6235 services_controller.go:360] Finished syncing service machine-api-operator-webhook on namespace openshift-machine-api for network=default : 2.387968ms\\\\nI1008 14:01:32.909851 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909853 6235 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:01:32.910156 6235 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.274715 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.287285 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.298021 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.308407 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.320949 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.357834 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.357863 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.357871 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.357885 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.357895 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.460228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.460277 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.460292 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.460309 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.460321 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.563257 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.564165 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.564249 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.564498 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.564533 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.667089 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.667116 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.667124 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.667136 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.667146 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.728962 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.729048 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:47 crc kubenswrapper[4789]: E1008 14:01:47.729218 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:47 crc kubenswrapper[4789]: E1008 14:01:47.729353 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.770484 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.770726 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.770819 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.770905 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.771040 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.873612 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.873656 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.873671 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.873692 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.873710 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.976316 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.976372 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.976381 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.976398 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:47 crc kubenswrapper[4789]: I1008 14:01:47.976408 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:47Z","lastTransitionTime":"2025-10-08T14:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.078780 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.078843 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.078862 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.078890 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.078910 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:48Z","lastTransitionTime":"2025-10-08T14:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.127713 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/2.log" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.128884 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/1.log" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.133827 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4" exitCode=1 Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.133892 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.133943 4789 scope.go:117] "RemoveContainer" containerID="5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.135192 4789 scope.go:117] "RemoveContainer" containerID="38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4" Oct 08 14:01:48 crc kubenswrapper[4789]: E1008 14:01:48.135593 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.159162 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.180683 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.182200 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.182268 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.182291 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.182324 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.182348 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:48Z","lastTransitionTime":"2025-10-08T14:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.200173 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.216456 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.230535 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.249408 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.265941 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.279665 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.285535 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.285582 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.285594 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.285614 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.285627 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:48Z","lastTransitionTime":"2025-10-08T14:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.295489 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.322806 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a2590c8d2befb59fdf95029f9ae14dd8f019d4a3713b0b9194e8e4d2ecd894c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:33Z\\\",\\\"message\\\":\\\"dler 7\\\\nI1008 14:01:32.909667 6235 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909660 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}\\\\nI1008 14:01:32.909711 6235 services_controller.go:360] Finished syncing service downloads on namespace openshift-console for network=default : 2.046028ms\\\\nI1008 14:01:32.909737 6235 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1008 14:01:32.909737 6235 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 14:01:32.909784 6235 factory.go:656] Stopping watch factory\\\\nI1008 14:01:32.909805 6235 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:01:32.909785 6235 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-webhook\\\\\\\"}\\\\nI1008 14:01:32.909874 6235 services_controller.go:360] Finished syncing service machine-api-operator-webhook on namespace openshift-machine-api for network=default : 2.387968ms\\\\nI1008 14:01:32.909851 6235 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 14:01:32.909853 6235 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:01:32.910156 6235 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.333189 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.343556 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.354835 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.365867 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.381831 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:48Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.387620 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.387738 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.387798 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.387880 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.387939 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:48Z","lastTransitionTime":"2025-10-08T14:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.491527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.491578 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.491588 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.491605 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.491618 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:48Z","lastTransitionTime":"2025-10-08T14:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.598286 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.598356 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.598381 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.598413 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.598474 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:48Z","lastTransitionTime":"2025-10-08T14:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.700670 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.700722 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.700731 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.700745 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.700756 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:48Z","lastTransitionTime":"2025-10-08T14:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.729921 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.729934 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:48 crc kubenswrapper[4789]: E1008 14:01:48.730145 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:48 crc kubenswrapper[4789]: E1008 14:01:48.730217 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.805161 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.805218 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.805257 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.805283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.805298 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:48Z","lastTransitionTime":"2025-10-08T14:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.908850 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.908923 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.908940 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.909058 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:48 crc kubenswrapper[4789]: I1008 14:01:48.909078 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:48Z","lastTransitionTime":"2025-10-08T14:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.012782 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.012863 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.012889 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.012920 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.012942 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.115787 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.115852 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.115868 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.115898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.115930 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.140277 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/2.log" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.145442 4789 scope.go:117] "RemoveContainer" containerID="38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4" Oct 08 14:01:49 crc kubenswrapper[4789]: E1008 14:01:49.145847 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.165502 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.187448 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.205585 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.219440 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.219516 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.219535 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.219563 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.219583 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.225354 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.248412 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.273112 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.298223 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.318616 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.323632 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.323675 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.323687 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.323704 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.323720 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.334652 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.358745 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.380411 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.401794 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.425581 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.427245 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.427304 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.427322 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.427348 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.427366 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.454528 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.470744 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:49Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.530578 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.530663 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.530689 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.530725 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.530752 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.634243 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.634328 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.634350 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.634378 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.634396 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.729280 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.729280 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:49 crc kubenswrapper[4789]: E1008 14:01:49.730261 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:49 crc kubenswrapper[4789]: E1008 14:01:49.730266 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.737927 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.738211 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.738414 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.738571 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.738715 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.842163 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.842273 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.842302 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.842336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.842360 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.946318 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.946391 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.946405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.946424 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:49 crc kubenswrapper[4789]: I1008 14:01:49.946436 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:49Z","lastTransitionTime":"2025-10-08T14:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.049828 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.049898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.049917 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.049943 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.049961 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.152577 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.152613 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.152624 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.152641 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.152654 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.255707 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.255753 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.255765 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.255782 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.255795 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.358965 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.359077 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.359097 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.359127 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.359147 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.462685 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.462774 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.462794 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.462827 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.462846 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.565750 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.565800 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.565818 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.565843 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.565861 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.669374 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.669441 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.669458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.669485 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.669504 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.697309 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:50 crc kubenswrapper[4789]: E1008 14:01:50.697586 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:50 crc kubenswrapper[4789]: E1008 14:01:50.697740 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs podName:765705a4-a303-4281-9677-5d0769f6d157 nodeName:}" failed. No retries permitted until 2025-10-08 14:02:06.697701282 +0000 UTC m=+66.604448814 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs") pod "network-metrics-daemon-b269s" (UID: "765705a4-a303-4281-9677-5d0769f6d157") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.729143 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.729198 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:50 crc kubenswrapper[4789]: E1008 14:01:50.729436 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:50 crc kubenswrapper[4789]: E1008 14:01:50.729608 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.750691 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.769245 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.771625 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.771822 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.771926 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.772134 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.772253 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.794135 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.811626 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.865505 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.875058 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.875129 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.875147 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.875174 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.875191 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.885030 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.904614 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.918648 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.937940 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.968281 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.981655 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.981719 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.981740 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.981783 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.981888 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:50Z","lastTransitionTime":"2025-10-08T14:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:50 crc kubenswrapper[4789]: I1008 14:01:50.992019 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:50Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.014835 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:51Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.032475 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:51Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.050144 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:51Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.064751 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:51Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.086149 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.086210 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.086228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.086261 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.086281 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:51Z","lastTransitionTime":"2025-10-08T14:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.189341 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.189376 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.189384 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.189399 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.189409 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:51Z","lastTransitionTime":"2025-10-08T14:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.293033 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.293085 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.293099 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.293119 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.293132 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:51Z","lastTransitionTime":"2025-10-08T14:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.395933 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.396042 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.396055 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.396072 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.396085 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:51Z","lastTransitionTime":"2025-10-08T14:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.498800 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.498853 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.498866 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.498885 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.498898 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:51Z","lastTransitionTime":"2025-10-08T14:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.601583 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.601624 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.601635 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.601653 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.601666 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:51Z","lastTransitionTime":"2025-10-08T14:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.704382 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.704414 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.704431 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.704448 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.704459 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:51Z","lastTransitionTime":"2025-10-08T14:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.728951 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.728951 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:51 crc kubenswrapper[4789]: E1008 14:01:51.729377 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:51 crc kubenswrapper[4789]: E1008 14:01:51.729446 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.806713 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.806783 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.806807 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.806837 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.806863 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:51Z","lastTransitionTime":"2025-10-08T14:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.909469 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.909499 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.909509 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.909525 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:51 crc kubenswrapper[4789]: I1008 14:01:51.909536 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:51Z","lastTransitionTime":"2025-10-08T14:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.011438 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.011703 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.011826 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.011946 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.012120 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.115001 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.115037 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.115046 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.115060 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.115074 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.217690 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.218019 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.218097 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.218180 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.218244 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.320452 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.320498 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.320509 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.320524 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.320535 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.423240 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.423289 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.423302 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.423320 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.423333 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.526220 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.526653 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.526796 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.526909 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.527055 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.630236 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.630300 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.630316 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.630345 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.630363 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.729886 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:52 crc kubenswrapper[4789]: E1008 14:01:52.730569 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.729940 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:52 crc kubenswrapper[4789]: E1008 14:01:52.730763 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.732886 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.733076 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.733162 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.733238 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.733301 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.836792 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.836917 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.836945 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.837025 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.837057 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.940337 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.940826 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.940929 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.941067 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:52 crc kubenswrapper[4789]: I1008 14:01:52.941212 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:52Z","lastTransitionTime":"2025-10-08T14:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.044300 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.044362 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.044380 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.044407 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.044429 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.148892 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.148951 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.148962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.148981 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.149014 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.252539 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.252589 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.252602 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.252622 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.252636 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.357773 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.357849 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.357868 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.357897 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.357922 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.461566 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.461626 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.461640 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.461658 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.461671 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.564804 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.564902 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.564932 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.564966 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.565028 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.637831 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.638081 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.638194 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638330 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638334 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:02:25.6382727 +0000 UTC m=+85.545020222 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638486 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:02:25.638462575 +0000 UTC m=+85.545210227 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638495 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638555 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638581 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.638540 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638673 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 14:02:25.63864249 +0000 UTC m=+85.545390012 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638672 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.638733 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638772 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:02:25.638747053 +0000 UTC m=+85.545494745 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638923 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638951 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.638969 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.639084 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 14:02:25.639054352 +0000 UTC m=+85.545801994 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.669479 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.669572 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.669591 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.669621 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.669641 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.729880 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.730050 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.730101 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:53 crc kubenswrapper[4789]: E1008 14:01:53.730285 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.773191 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.773253 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.773267 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.773291 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.773307 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.877485 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.877600 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.877623 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.877655 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.877679 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.983204 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.983273 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.983293 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.983323 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:53 crc kubenswrapper[4789]: I1008 14:01:53.983344 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:53Z","lastTransitionTime":"2025-10-08T14:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.086357 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.086433 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.086459 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.086492 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.086516 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:54Z","lastTransitionTime":"2025-10-08T14:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.190940 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.191272 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.191449 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.191767 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.191880 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:54Z","lastTransitionTime":"2025-10-08T14:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.295344 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.295398 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.295411 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.295432 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.295446 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:54Z","lastTransitionTime":"2025-10-08T14:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.398646 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.398696 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.398708 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.398726 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.398737 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:54Z","lastTransitionTime":"2025-10-08T14:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.501380 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.501417 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.501426 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.501442 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.501453 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:54Z","lastTransitionTime":"2025-10-08T14:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.603266 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.603301 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.603310 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.603326 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.603336 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:54Z","lastTransitionTime":"2025-10-08T14:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.707279 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.707333 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.707345 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.707366 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.707382 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:54Z","lastTransitionTime":"2025-10-08T14:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.729245 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:54 crc kubenswrapper[4789]: E1008 14:01:54.729684 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.729912 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:54 crc kubenswrapper[4789]: E1008 14:01:54.730277 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.809242 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.809305 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.809324 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.809354 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.809373 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:54Z","lastTransitionTime":"2025-10-08T14:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.912309 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.912784 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.913040 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.913280 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:54 crc kubenswrapper[4789]: I1008 14:01:54.913488 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:54Z","lastTransitionTime":"2025-10-08T14:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.016300 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.016773 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.016898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.017094 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.017242 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.120272 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.120558 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.120644 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.120736 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.120815 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.223843 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.223907 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.223922 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.223947 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.223965 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.326678 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.326730 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.326741 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.326761 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.326775 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.429795 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.429885 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.429901 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.429922 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.429935 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.532758 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.532807 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.532819 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.532840 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.532852 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.635783 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.635837 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.635849 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.635868 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.635882 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.729432 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:55 crc kubenswrapper[4789]: E1008 14:01:55.729595 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.729465 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:55 crc kubenswrapper[4789]: E1008 14:01:55.730044 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.738489 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.739071 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.739092 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.739113 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.739125 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.841883 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.841925 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.841939 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.841962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.841977 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.944242 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.944286 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.944296 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.944311 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:55 crc kubenswrapper[4789]: I1008 14:01:55.944320 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:55Z","lastTransitionTime":"2025-10-08T14:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.047251 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.047310 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.047329 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.047355 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.047381 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.095317 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.108908 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.118843 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.133663 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.150494 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.151413 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.151452 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.151465 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.151484 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.151497 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.173602 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.188106 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.204156 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.222310 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.240310 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.252726 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.254405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.254463 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.254475 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.254493 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.254506 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.279296 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.298284 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.316665 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.334022 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.358962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.359025 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.359037 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.359058 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.359110 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.370763 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.385884 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:56Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.462163 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.462236 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.462261 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.462280 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.462293 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.566438 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.566476 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.566490 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.566506 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.566517 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.670727 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.670792 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.670809 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.670836 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.670855 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.729410 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.729572 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:56 crc kubenswrapper[4789]: E1008 14:01:56.729605 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:56 crc kubenswrapper[4789]: E1008 14:01:56.729811 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.773439 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.773484 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.773501 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.773523 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.773542 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.876916 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.876983 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.877045 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.877074 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.877094 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.979518 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.979583 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.979601 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.979628 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:56 crc kubenswrapper[4789]: I1008 14:01:56.979651 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:56Z","lastTransitionTime":"2025-10-08T14:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.065771 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.065834 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.065854 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.065880 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.065897 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: E1008 14:01:57.088370 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:57Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.094573 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.094682 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.094752 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.094783 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.094844 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: E1008 14:01:57.115204 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:57Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.121539 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.121589 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.121605 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.121628 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.121647 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: E1008 14:01:57.138533 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:57Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.146771 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.146822 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.146838 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.146862 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.146879 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: E1008 14:01:57.160336 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:57Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.164728 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.164857 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.164877 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.164903 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.165142 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: E1008 14:01:57.184778 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:57Z is after 2025-08-24T17:21:41Z" Oct 08 14:01:57 crc kubenswrapper[4789]: E1008 14:01:57.184916 4789 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.186680 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.186702 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.186709 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.186720 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.186752 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.289013 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.289043 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.289052 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.289065 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.289089 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.392065 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.392113 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.392131 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.392156 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.392175 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.495234 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.495270 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.495282 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.495300 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.495311 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.598527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.598593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.598610 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.598635 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.598656 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.701685 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.701722 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.701731 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.701746 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.701755 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.729475 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.729598 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:57 crc kubenswrapper[4789]: E1008 14:01:57.729624 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:57 crc kubenswrapper[4789]: E1008 14:01:57.729856 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.804368 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.804446 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.804462 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.804483 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.804496 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.907217 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.907313 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.907338 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.907369 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:57 crc kubenswrapper[4789]: I1008 14:01:57.907392 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:57Z","lastTransitionTime":"2025-10-08T14:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.009964 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.010025 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.010036 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.010053 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.010065 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.113957 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.114078 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.114105 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.114128 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.114151 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.216409 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.216437 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.216447 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.216463 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.216474 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.319448 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.319499 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.319517 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.319542 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.319560 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.422527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.422796 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.422868 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.422937 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.423032 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.525805 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.525875 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.525898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.525932 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.525955 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.634042 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.634115 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.634138 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.634167 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.634190 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.730155 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:01:58 crc kubenswrapper[4789]: E1008 14:01:58.730396 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.730572 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:01:58 crc kubenswrapper[4789]: E1008 14:01:58.730867 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.737128 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.737456 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.737623 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.737784 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.737924 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.841480 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.841545 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.841562 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.841588 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.841621 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.944112 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.944431 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.944564 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.944672 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:58 crc kubenswrapper[4789]: I1008 14:01:58.944781 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:58Z","lastTransitionTime":"2025-10-08T14:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.048320 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.048378 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.048396 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.048433 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.048451 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.150876 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.151175 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.151239 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.151317 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.151381 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.254643 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.254688 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.254697 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.254714 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.254724 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.358158 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.358208 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.358216 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.358233 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.358245 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.461243 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.461303 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.461320 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.461347 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.461366 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.564538 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.564951 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.565191 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.565387 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.565572 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.686722 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.686772 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.686782 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.686803 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.686825 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.729727 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.729867 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:01:59 crc kubenswrapper[4789]: E1008 14:01:59.729971 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:01:59 crc kubenswrapper[4789]: E1008 14:01:59.730314 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.789189 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.789261 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.789283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.789310 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.789328 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.892109 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.892186 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.892198 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.892213 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.892222 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.994540 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.994590 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.994611 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.994640 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:01:59 crc kubenswrapper[4789]: I1008 14:01:59.994661 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:01:59Z","lastTransitionTime":"2025-10-08T14:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.097608 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.097643 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.097659 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.097678 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.097692 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:00Z","lastTransitionTime":"2025-10-08T14:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.199673 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.199815 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.199828 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.199843 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.199854 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:00Z","lastTransitionTime":"2025-10-08T14:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.303113 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.303196 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.303222 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.303261 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.303287 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:00Z","lastTransitionTime":"2025-10-08T14:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.406911 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.407346 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.407515 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.407680 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.407843 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:00Z","lastTransitionTime":"2025-10-08T14:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.511620 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.511682 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.511700 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.511726 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.511744 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:00Z","lastTransitionTime":"2025-10-08T14:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.615976 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.616473 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.616676 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.616882 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.617113 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:00Z","lastTransitionTime":"2025-10-08T14:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.720336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.720382 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.720396 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.720417 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.720430 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:00Z","lastTransitionTime":"2025-10-08T14:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.729639 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.729742 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:00 crc kubenswrapper[4789]: E1008 14:02:00.729891 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:00 crc kubenswrapper[4789]: E1008 14:02:00.729947 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.745170 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.761894 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.779243 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.801662 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.817522 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.822889 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.822953 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.822971 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.823032 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.823058 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:00Z","lastTransitionTime":"2025-10-08T14:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.832664 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e0199c9-c939-49f7-991b-b9c6fef2f435\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.852294 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.867766 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.884690 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.905874 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.917112 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.927576 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.927624 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.927637 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.927656 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.927668 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:00Z","lastTransitionTime":"2025-10-08T14:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.928826 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.971222 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:00 crc kubenswrapper[4789]: I1008 14:02:00.987412 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.000102 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:00Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.011491 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:01Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.030231 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.030291 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.030309 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.030378 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.030399 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.134733 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.134777 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.134786 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.134801 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.134811 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.237158 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.237214 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.237229 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.237251 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.237267 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.340088 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.340163 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.340184 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.340216 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.340238 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.443312 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.443569 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.443656 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.443727 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.443794 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.546308 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.546378 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.546395 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.546421 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.546438 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.650184 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.650231 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.650247 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.650275 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.650293 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.730102 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:01 crc kubenswrapper[4789]: E1008 14:02:01.730329 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.730102 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:01 crc kubenswrapper[4789]: E1008 14:02:01.730887 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.753569 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.753899 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.754134 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.754307 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.754767 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.858479 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.858538 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.858556 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.858579 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.858596 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.961613 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.962822 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.962973 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.963156 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:01 crc kubenswrapper[4789]: I1008 14:02:01.963308 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:01Z","lastTransitionTime":"2025-10-08T14:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.066270 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.066670 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.066812 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.066962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.067255 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.169915 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.170237 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.170408 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.170543 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.170687 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.274152 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.274434 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.274532 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.274630 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.274715 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.377849 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.377908 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.377925 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.377955 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.377975 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.480870 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.481058 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.481087 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.481116 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.481137 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.584544 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.584924 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.584942 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.584967 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.584986 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.688324 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.688414 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.688438 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.688481 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.688506 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.729542 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:02 crc kubenswrapper[4789]: E1008 14:02:02.729706 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.729567 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:02 crc kubenswrapper[4789]: E1008 14:02:02.730520 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.791645 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.791706 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.791726 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.791760 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.791779 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.895250 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.895312 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.895330 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.895360 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.895380 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.998515 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.998586 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.998602 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.998627 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:02 crc kubenswrapper[4789]: I1008 14:02:02.998646 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:02Z","lastTransitionTime":"2025-10-08T14:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.102776 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.102865 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.102894 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.102929 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.102950 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:03Z","lastTransitionTime":"2025-10-08T14:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.206047 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.206112 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.206132 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.206157 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.206170 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:03Z","lastTransitionTime":"2025-10-08T14:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.308866 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.308928 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.308946 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.308970 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.309017 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:03Z","lastTransitionTime":"2025-10-08T14:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.412613 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.412670 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.412686 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.412711 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.412729 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:03Z","lastTransitionTime":"2025-10-08T14:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.516228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.516736 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.516880 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.517082 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.517224 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:03Z","lastTransitionTime":"2025-10-08T14:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.620898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.620941 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.620953 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.620973 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.621009 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:03Z","lastTransitionTime":"2025-10-08T14:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.724582 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.724620 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.724631 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.724647 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.724658 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:03Z","lastTransitionTime":"2025-10-08T14:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.730775 4789 scope.go:117] "RemoveContainer" containerID="38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4" Oct 08 14:02:03 crc kubenswrapper[4789]: E1008 14:02:03.730979 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.731190 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:03 crc kubenswrapper[4789]: E1008 14:02:03.731266 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.731397 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:03 crc kubenswrapper[4789]: E1008 14:02:03.731453 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.828443 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.828837 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.829027 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.829204 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.829346 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:03Z","lastTransitionTime":"2025-10-08T14:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.932445 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.932500 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.932517 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.932540 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:03 crc kubenswrapper[4789]: I1008 14:02:03.932561 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:03Z","lastTransitionTime":"2025-10-08T14:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.035951 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.036040 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.036058 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.036084 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.036104 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.140410 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.140540 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.140615 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.140663 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.140739 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.243602 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.243653 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.243665 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.243685 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.243699 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.347239 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.347332 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.347374 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.347397 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.347411 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.451300 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.451376 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.451401 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.451430 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.451453 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.554228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.554279 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.554292 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.554310 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.554322 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.656741 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.656959 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.657067 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.657156 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.657223 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.729684 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.730026 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:04 crc kubenswrapper[4789]: E1008 14:02:04.730129 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:04 crc kubenswrapper[4789]: E1008 14:02:04.730286 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.760394 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.760617 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.760741 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.760842 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.760933 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.862904 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.862964 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.862979 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.863018 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.863032 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.965970 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.966032 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.966043 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.966066 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:04 crc kubenswrapper[4789]: I1008 14:02:04.966080 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:04Z","lastTransitionTime":"2025-10-08T14:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.069724 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.070242 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.070348 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.070428 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.070504 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:05Z","lastTransitionTime":"2025-10-08T14:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.174106 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.174157 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.174169 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.174189 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.174201 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:05Z","lastTransitionTime":"2025-10-08T14:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.275938 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.275964 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.275972 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.276001 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.276010 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:05Z","lastTransitionTime":"2025-10-08T14:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.381226 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.381612 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.381793 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.381966 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.382144 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:05Z","lastTransitionTime":"2025-10-08T14:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.485219 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.485282 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.485303 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.485334 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.485351 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:05Z","lastTransitionTime":"2025-10-08T14:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.587570 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.587624 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.587636 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.587654 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.587665 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:05Z","lastTransitionTime":"2025-10-08T14:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.691334 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.691407 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.691432 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.691465 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.691483 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:05Z","lastTransitionTime":"2025-10-08T14:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.729461 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:05 crc kubenswrapper[4789]: E1008 14:02:05.729885 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.729468 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:05 crc kubenswrapper[4789]: E1008 14:02:05.730164 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.795956 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.796093 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.796122 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.796176 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.796203 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:05Z","lastTransitionTime":"2025-10-08T14:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.898555 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.898628 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.898651 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.898679 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:05 crc kubenswrapper[4789]: I1008 14:02:05.898701 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:05Z","lastTransitionTime":"2025-10-08T14:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.022865 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.022921 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.022944 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.022972 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.023022 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.125894 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.125956 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.125978 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.126037 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.126061 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.228134 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.228264 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.228283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.228319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.228340 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.331076 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.331131 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.331146 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.331166 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.331180 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.433820 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.433884 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.433896 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.433912 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.433925 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.536125 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.536169 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.536180 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.536198 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.536210 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.638939 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.639033 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.639052 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.639079 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.639097 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.729896 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.729897 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:06 crc kubenswrapper[4789]: E1008 14:02:06.730047 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:06 crc kubenswrapper[4789]: E1008 14:02:06.730169 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.741266 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.741352 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.741372 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.741393 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.741410 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.760841 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:06 crc kubenswrapper[4789]: E1008 14:02:06.761219 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:02:06 crc kubenswrapper[4789]: E1008 14:02:06.761363 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs podName:765705a4-a303-4281-9677-5d0769f6d157 nodeName:}" failed. No retries permitted until 2025-10-08 14:02:38.76134681 +0000 UTC m=+98.668094302 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs") pod "network-metrics-daemon-b269s" (UID: "765705a4-a303-4281-9677-5d0769f6d157") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.844096 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.844325 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.844391 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.844472 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.844560 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.946539 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.946575 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.946583 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.946599 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:06 crc kubenswrapper[4789]: I1008 14:02:06.946609 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:06Z","lastTransitionTime":"2025-10-08T14:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.049546 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.049605 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.049625 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.049652 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.049674 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.151634 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.151673 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.151684 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.151700 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.151714 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.254516 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.254779 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.254847 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.254928 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.255043 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.357630 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.357692 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.357714 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.357743 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.357766 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.460238 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.460280 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.460291 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.460306 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.460315 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.563187 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.563225 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.563233 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.563247 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.563257 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.586279 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.586311 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.586319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.586333 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.586343 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: E1008 14:02:07.607620 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:07Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.611722 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.611761 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.611774 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.611791 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.611804 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: E1008 14:02:07.631857 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:07Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.634957 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.635025 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.635044 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.635065 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.635082 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: E1008 14:02:07.649200 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:07Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.652916 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.652978 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.653017 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.653038 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.653052 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: E1008 14:02:07.667161 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:07Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.674880 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.674938 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.674955 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.674978 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.675030 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: E1008 14:02:07.689302 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:07Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:07 crc kubenswrapper[4789]: E1008 14:02:07.689417 4789 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.691115 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.691142 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.691152 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.691194 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.691203 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.729599 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.729706 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:07 crc kubenswrapper[4789]: E1008 14:02:07.729800 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:07 crc kubenswrapper[4789]: E1008 14:02:07.729978 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.793723 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.793827 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.793844 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.793860 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.793869 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.897019 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.897049 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.897086 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.897102 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.897114 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.999271 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.999317 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.999328 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.999346 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:07 crc kubenswrapper[4789]: I1008 14:02:07.999358 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:07Z","lastTransitionTime":"2025-10-08T14:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.101446 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.101496 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.101512 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.101535 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.101553 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:08Z","lastTransitionTime":"2025-10-08T14:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.203857 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.203885 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.203893 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.203907 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.203916 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:08Z","lastTransitionTime":"2025-10-08T14:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.306208 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.306248 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.306258 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.306274 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.306298 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:08Z","lastTransitionTime":"2025-10-08T14:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.407832 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.407867 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.407875 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.407889 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.407900 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:08Z","lastTransitionTime":"2025-10-08T14:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.510350 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.510385 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.510396 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.510411 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.510420 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:08Z","lastTransitionTime":"2025-10-08T14:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.612556 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.612588 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.612596 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.612610 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.612621 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:08Z","lastTransitionTime":"2025-10-08T14:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.715047 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.715079 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.715090 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.715106 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.715116 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:08Z","lastTransitionTime":"2025-10-08T14:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.729562 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.729707 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:08 crc kubenswrapper[4789]: E1008 14:02:08.729929 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:08 crc kubenswrapper[4789]: E1008 14:02:08.730119 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.817870 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.817923 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.817932 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.817961 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.817971 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:08Z","lastTransitionTime":"2025-10-08T14:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.920755 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.920814 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.920831 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.920856 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:08 crc kubenswrapper[4789]: I1008 14:02:08.920873 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:08Z","lastTransitionTime":"2025-10-08T14:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.023336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.023373 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.023383 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.023416 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.023426 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.126361 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.126406 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.126416 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.126432 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.126443 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.229004 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.229057 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.229115 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.229136 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.229156 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.332470 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.332520 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.332534 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.332557 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.332569 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.435776 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.435818 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.435829 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.435843 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.435853 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.538265 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.538320 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.538337 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.538359 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.538375 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.640953 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.641073 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.641095 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.641121 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.641147 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.738382 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.738441 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:09 crc kubenswrapper[4789]: E1008 14:02:09.738633 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:09 crc kubenswrapper[4789]: E1008 14:02:09.739119 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.743866 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.744034 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.744067 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.744097 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.744122 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.846602 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.846650 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.846661 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.846678 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.846690 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.949451 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.949499 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.949510 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.949527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:09 crc kubenswrapper[4789]: I1008 14:02:09.949538 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:09Z","lastTransitionTime":"2025-10-08T14:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.051735 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.051771 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.051782 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.051796 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.051807 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.154304 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.154343 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.154351 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.154366 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.154375 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.222828 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/0.log" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.222909 4789 generic.go:334] "Generic (PLEG): container finished" podID="4aee6ee4-49ed-4135-8bb0-003eb4f70f05" containerID="b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29" exitCode=1 Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.222953 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5428v" event={"ID":"4aee6ee4-49ed-4135-8bb0-003eb4f70f05","Type":"ContainerDied","Data":"b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.223539 4789 scope.go:117] "RemoveContainer" containerID="b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.234739 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.250624 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.257489 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.257554 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.257579 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.257607 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.257627 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.264631 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:09Z\\\",\\\"message\\\":\\\"2025-10-08T14:01:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c\\\\n2025-10-08T14:01:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c to /host/opt/cni/bin/\\\\n2025-10-08T14:01:24Z [verbose] multus-daemon started\\\\n2025-10-08T14:01:24Z [verbose] Readiness Indicator file check\\\\n2025-10-08T14:02:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.282521 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e0199c9-c939-49f7-991b-b9c6fef2f435\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.307960 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.323892 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.339631 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.360358 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.360404 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.360419 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.360439 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.360454 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.366307 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.379354 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.391133 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.401038 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.414468 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.426183 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.436359 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.446125 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.455790 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.462381 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.462406 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.462415 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.462428 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.462436 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.564505 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.564536 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.564544 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.564556 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.564565 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.666978 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.667057 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.667070 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.667090 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.667104 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.730289 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.730297 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:10 crc kubenswrapper[4789]: E1008 14:02:10.730519 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:10 crc kubenswrapper[4789]: E1008 14:02:10.730573 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.739792 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.748369 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.760367 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:09Z\\\",\\\"message\\\":\\\"2025-10-08T14:01:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c\\\\n2025-10-08T14:01:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c to /host/opt/cni/bin/\\\\n2025-10-08T14:01:24Z [verbose] multus-daemon started\\\\n2025-10-08T14:01:24Z [verbose] Readiness Indicator file check\\\\n2025-10-08T14:02:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.769357 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.769508 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.769582 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.769650 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.769714 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.777380 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e0199c9-c939-49f7-991b-b9c6fef2f435\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.790216 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.802844 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.815789 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.827370 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.846999 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.857099 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.866303 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.872295 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.872318 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.872328 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.872342 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.872352 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.882301 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.891402 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.904770 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.916520 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.927164 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.939366 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:10Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.975003 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.975049 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.975061 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.975078 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:10 crc kubenswrapper[4789]: I1008 14:02:10.975090 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:10Z","lastTransitionTime":"2025-10-08T14:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.077290 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.077337 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.077346 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.077360 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.077369 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.179728 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.179772 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.179780 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.179795 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.179806 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.227732 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/0.log" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.227828 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5428v" event={"ID":"4aee6ee4-49ed-4135-8bb0-003eb4f70f05","Type":"ContainerStarted","Data":"25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.237879 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.250323 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.262070 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:09Z\\\",\\\"message\\\":\\\"2025-10-08T14:01:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c\\\\n2025-10-08T14:01:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c to /host/opt/cni/bin/\\\\n2025-10-08T14:01:24Z [verbose] multus-daemon started\\\\n2025-10-08T14:01:24Z [verbose] Readiness Indicator file check\\\\n2025-10-08T14:02:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.274331 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e0199c9-c939-49f7-991b-b9c6fef2f435\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.281968 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.282041 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.282052 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.282069 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.282081 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.286000 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.296648 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.308662 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.318519 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.328374 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.338406 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"390b82b2-610e-44c0-b509-c2e26db021e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb4658ef6d767e0b260c91435cffde1ce49a2ef282408f85c68c7b459323f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.348033 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.360046 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.378698 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.383878 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.383927 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.383938 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.383954 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.383966 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.392868 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.405365 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.414582 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.424813 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:11Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.486474 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.486504 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.486511 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.486524 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.486533 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.588912 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.589012 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.589027 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.589045 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.589057 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.690949 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.691084 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.691140 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.691163 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.691184 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.729161 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.729299 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:11 crc kubenswrapper[4789]: E1008 14:02:11.729323 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:11 crc kubenswrapper[4789]: E1008 14:02:11.729490 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.793403 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.793454 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.793466 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.793484 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.793498 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.896111 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.896166 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.896180 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.896199 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.896211 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.998326 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.998388 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.998405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.998434 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:11 crc kubenswrapper[4789]: I1008 14:02:11.998452 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:11Z","lastTransitionTime":"2025-10-08T14:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.100760 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.100802 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.100818 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.100843 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.100859 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:12Z","lastTransitionTime":"2025-10-08T14:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.203417 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.203449 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.203457 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.203470 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.203479 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:12Z","lastTransitionTime":"2025-10-08T14:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.306100 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.306141 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.306149 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.306163 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.306173 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:12Z","lastTransitionTime":"2025-10-08T14:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.408251 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.408294 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.408302 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.408316 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.408325 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:12Z","lastTransitionTime":"2025-10-08T14:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.510943 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.511022 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.511036 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.511058 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.511069 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:12Z","lastTransitionTime":"2025-10-08T14:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.613807 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.613865 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.613881 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.613901 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.613913 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:12Z","lastTransitionTime":"2025-10-08T14:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.716483 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.716789 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.716956 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.717225 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.717384 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:12Z","lastTransitionTime":"2025-10-08T14:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.729545 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.729601 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:12 crc kubenswrapper[4789]: E1008 14:02:12.729791 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:12 crc kubenswrapper[4789]: E1008 14:02:12.729809 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.819712 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.819760 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.819770 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.819787 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.819798 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:12Z","lastTransitionTime":"2025-10-08T14:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.922526 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.922598 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.922616 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.922645 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:12 crc kubenswrapper[4789]: I1008 14:02:12.922663 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:12Z","lastTransitionTime":"2025-10-08T14:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.024822 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.024964 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.025016 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.025046 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.025065 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.128302 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.128361 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.128379 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.128402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.128419 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.231388 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.231444 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.231457 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.231478 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.231491 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.334794 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.334851 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.334867 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.334895 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.334911 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.437814 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.437884 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.437906 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.437934 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.437955 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.541178 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.541255 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.541279 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.541307 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.541334 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.644993 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.645159 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.645184 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.645222 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.645248 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.730134 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.730134 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:13 crc kubenswrapper[4789]: E1008 14:02:13.730367 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:13 crc kubenswrapper[4789]: E1008 14:02:13.730452 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.747853 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.747913 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.747929 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.747957 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.747976 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.851377 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.851657 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.851670 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.851689 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.851700 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.955141 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.955222 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.955244 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.955276 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:13 crc kubenswrapper[4789]: I1008 14:02:13.955302 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:13Z","lastTransitionTime":"2025-10-08T14:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.059227 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.059608 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.059634 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.059661 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.059680 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:14Z","lastTransitionTime":"2025-10-08T14:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.164033 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.164391 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.164549 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.164703 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.164836 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:14Z","lastTransitionTime":"2025-10-08T14:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.267922 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.267981 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.268017 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.268045 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.268060 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:14Z","lastTransitionTime":"2025-10-08T14:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.371438 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.371659 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.371686 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.372493 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.372686 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:14Z","lastTransitionTime":"2025-10-08T14:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.476711 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.476772 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.476786 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.476803 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.476814 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:14Z","lastTransitionTime":"2025-10-08T14:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.580368 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.580452 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.580475 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.580505 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.580528 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:14Z","lastTransitionTime":"2025-10-08T14:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.686461 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.686529 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.686549 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.686612 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.686634 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:14Z","lastTransitionTime":"2025-10-08T14:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.730161 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.730189 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:14 crc kubenswrapper[4789]: E1008 14:02:14.730368 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:14 crc kubenswrapper[4789]: E1008 14:02:14.730621 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.795272 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.795384 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.795418 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.795474 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.795500 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:14Z","lastTransitionTime":"2025-10-08T14:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.900113 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.900168 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.900187 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.900215 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:14 crc kubenswrapper[4789]: I1008 14:02:14.900232 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:14Z","lastTransitionTime":"2025-10-08T14:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.004939 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.005109 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.005134 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.005171 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.005193 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.108442 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.108497 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.108510 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.108527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.108537 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.212369 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.212478 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.212619 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.212710 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.212804 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.316882 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.316961 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.316988 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.317070 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.317116 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.420324 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.420387 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.420405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.420437 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.420457 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.524287 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.524713 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.524730 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.524748 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.524759 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.628413 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.628489 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.628514 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.628547 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.628571 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.729252 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.729259 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:15 crc kubenswrapper[4789]: E1008 14:02:15.729880 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:15 crc kubenswrapper[4789]: E1008 14:02:15.730027 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.731225 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.731306 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.731331 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.731360 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.731383 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.834245 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.834342 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.834361 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.834397 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.834428 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.937890 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.938240 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.938342 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.938430 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:15 crc kubenswrapper[4789]: I1008 14:02:15.938535 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:15Z","lastTransitionTime":"2025-10-08T14:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.041832 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.041895 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.041912 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.041941 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.041968 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.145808 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.146579 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.146628 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.146660 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.146678 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.250009 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.250054 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.250062 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.250079 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.250089 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.353429 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.353484 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.353497 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.353521 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.353537 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.456957 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.457038 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.457052 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.457072 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.457086 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.559589 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.559636 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.559647 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.559663 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.559676 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.662245 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.662301 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.662312 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.662330 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.662346 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.729331 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.729410 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:16 crc kubenswrapper[4789]: E1008 14:02:16.729550 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:16 crc kubenswrapper[4789]: E1008 14:02:16.729724 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.765526 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.765591 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.765606 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.765627 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.765643 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.867725 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.867770 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.867780 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.867796 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.867807 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.970389 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.970426 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.970439 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.970455 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:16 crc kubenswrapper[4789]: I1008 14:02:16.970465 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:16Z","lastTransitionTime":"2025-10-08T14:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.073324 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.073364 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.073375 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.073392 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.073404 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:17Z","lastTransitionTime":"2025-10-08T14:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.176115 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.176160 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.176172 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.176190 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.176204 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:17Z","lastTransitionTime":"2025-10-08T14:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.279528 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.279612 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.279634 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.279666 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.279686 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:17Z","lastTransitionTime":"2025-10-08T14:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.387257 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.387379 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.387407 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.387449 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.387486 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:17Z","lastTransitionTime":"2025-10-08T14:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.491509 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.491591 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.491607 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.491633 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.491701 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:17Z","lastTransitionTime":"2025-10-08T14:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.594963 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.595057 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.595071 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.595121 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.595136 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:17Z","lastTransitionTime":"2025-10-08T14:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.698086 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.698169 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.698188 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.698221 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.698244 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:17Z","lastTransitionTime":"2025-10-08T14:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.729310 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.729352 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:17 crc kubenswrapper[4789]: E1008 14:02:17.729457 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:17 crc kubenswrapper[4789]: E1008 14:02:17.729646 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.800570 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.800607 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.800616 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.800633 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.800644 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:17Z","lastTransitionTime":"2025-10-08T14:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.903559 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.903591 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.903599 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.903615 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:17 crc kubenswrapper[4789]: I1008 14:02:17.903624 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:17Z","lastTransitionTime":"2025-10-08T14:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.006354 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.006394 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.006403 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.006416 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.006425 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.037555 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.037598 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.037610 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.037629 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.037640 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: E1008 14:02:18.049574 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:18Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.053392 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.053431 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.053441 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.053458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.053467 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: E1008 14:02:18.065174 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:18Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.069673 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.069774 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.069818 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.069845 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.069861 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: E1008 14:02:18.087854 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:18Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.092139 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.092174 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.092184 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.092203 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.092217 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: E1008 14:02:18.104184 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:18Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.108513 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.108559 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.108569 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.108586 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.108596 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: E1008 14:02:18.119405 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:18Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:18 crc kubenswrapper[4789]: E1008 14:02:18.119520 4789 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.121204 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.121229 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.121237 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.121251 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.121260 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.224425 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.224492 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.224506 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.224528 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.224540 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.327711 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.327763 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.327778 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.327800 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.327812 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.430814 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.430881 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.430932 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.430967 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.430986 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.535288 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.535332 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.535342 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.535360 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.535371 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.639110 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.639186 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.639205 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.639236 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.639254 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.732401 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.733068 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:18 crc kubenswrapper[4789]: E1008 14:02:18.733230 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:18 crc kubenswrapper[4789]: E1008 14:02:18.733611 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.733637 4789 scope.go:117] "RemoveContainer" containerID="38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.742093 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.742139 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.742158 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.742180 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.742198 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.847459 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.847504 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.847514 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.847533 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.847546 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.951669 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.951723 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.951741 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.951766 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:18 crc kubenswrapper[4789]: I1008 14:02:18.951785 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:18Z","lastTransitionTime":"2025-10-08T14:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.054600 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.054643 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.054654 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.054671 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.054684 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.157753 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.157797 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.157809 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.157826 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.157836 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.257796 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/2.log" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.259370 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.259450 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.259465 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.259491 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.259519 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.260887 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.261451 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.276538 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"390b82b2-610e-44c0-b509-c2e26db021e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb4658ef6d767e0b260c91435cffde1ce49a2ef282408f85c68c7b459323f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.298560 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.316561 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.339075 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.355209 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.361575 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.361784 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.361875 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.361973 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.362087 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.372310 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.386000 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.400602 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.413103 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.427165 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.444354 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e0199c9-c939-49f7-991b-b9c6fef2f435\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.462837 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.464208 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.464243 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.464284 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.464303 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.464316 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.487774 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.503495 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.517891 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.538189 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.553206 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:09Z\\\",\\\"message\\\":\\\"2025-10-08T14:01:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c\\\\n2025-10-08T14:01:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c to /host/opt/cni/bin/\\\\n2025-10-08T14:01:24Z [verbose] multus-daemon started\\\\n2025-10-08T14:01:24Z [verbose] Readiness Indicator file check\\\\n2025-10-08T14:02:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:19Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.567659 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.567708 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.567725 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.567748 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.567767 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.670672 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.671159 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.671234 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.671309 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.671405 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.729374 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:19 crc kubenswrapper[4789]: E1008 14:02:19.729964 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.729714 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:19 crc kubenswrapper[4789]: E1008 14:02:19.730348 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.774512 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.774929 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.775033 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.775107 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.775180 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.878338 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.878380 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.878391 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.878408 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.878417 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.983299 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.983369 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.983387 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.983417 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:19 crc kubenswrapper[4789]: I1008 14:02:19.983436 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:19Z","lastTransitionTime":"2025-10-08T14:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.086465 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.086506 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.086517 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.086533 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.086545 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:20Z","lastTransitionTime":"2025-10-08T14:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.189574 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.189626 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.189642 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.189663 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.189679 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:20Z","lastTransitionTime":"2025-10-08T14:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.266115 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/3.log" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.267064 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/2.log" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.270651 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298" exitCode=1 Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.270700 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.270758 4789 scope.go:117] "RemoveContainer" containerID="38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.271809 4789 scope.go:117] "RemoveContainer" containerID="3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298" Oct 08 14:02:20 crc kubenswrapper[4789]: E1008 14:02:20.272100 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.296610 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.296648 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.296664 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.296682 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.296696 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:20Z","lastTransitionTime":"2025-10-08T14:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.302778 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:19Z\\\",\\\"message\\\":\\\"nNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:02:19.767861 6800 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:02:19.767873 6800 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 14:02:19.767890 6800 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 14:02:19.767902 6800 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 14:02:19.767911 6800 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 14:02:19.767927 6800 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:02:19.767944 6800 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 14:02:19.767957 6800 factory.go:656] Stopping watch factory\\\\nI1008 14:02:19.767964 6800 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 14:02:19.767978 6800 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 14:02:19.768004 6800 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 14:02:19.768012 6800 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 14:02:19.768342 6800 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:02:19.768387 6800 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:02:19.768524 6800 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.317137 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.331181 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"390b82b2-610e-44c0-b509-c2e26db021e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb4658ef6d767e0b260c91435cffde1ce49a2ef282408f85c68c7b459323f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.345109 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.357092 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.369541 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.383664 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.398656 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.400031 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.400078 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.400087 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.400104 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.400115 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:20Z","lastTransitionTime":"2025-10-08T14:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.411708 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.423838 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.433931 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.449198 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.467817 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:09Z\\\",\\\"message\\\":\\\"2025-10-08T14:01:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c\\\\n2025-10-08T14:01:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c to /host/opt/cni/bin/\\\\n2025-10-08T14:01:24Z [verbose] multus-daemon started\\\\n2025-10-08T14:01:24Z [verbose] Readiness Indicator file check\\\\n2025-10-08T14:02:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.484539 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e0199c9-c939-49f7-991b-b9c6fef2f435\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.500630 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.503121 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.503167 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.503184 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.503209 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.503231 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:20Z","lastTransitionTime":"2025-10-08T14:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.516751 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.533105 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.606184 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.606225 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.606242 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.606313 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.606332 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:20Z","lastTransitionTime":"2025-10-08T14:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.709541 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.709587 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.709598 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.709616 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.709630 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:20Z","lastTransitionTime":"2025-10-08T14:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.729859 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.729937 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:20 crc kubenswrapper[4789]: E1008 14:02:20.730606 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:20 crc kubenswrapper[4789]: E1008 14:02:20.730929 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.746287 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.757753 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.769750 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.790945 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.805425 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.813041 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.813092 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.813103 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.813124 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.813137 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:20Z","lastTransitionTime":"2025-10-08T14:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.819425 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.836736 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.850406 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.867557 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.883065 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.897371 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:09Z\\\",\\\"message\\\":\\\"2025-10-08T14:01:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c\\\\n2025-10-08T14:01:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c to /host/opt/cni/bin/\\\\n2025-10-08T14:01:24Z [verbose] multus-daemon started\\\\n2025-10-08T14:01:24Z [verbose] Readiness Indicator file check\\\\n2025-10-08T14:02:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.908897 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e0199c9-c939-49f7-991b-b9c6fef2f435\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.916100 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.916147 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.916158 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.916175 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.916185 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:20Z","lastTransitionTime":"2025-10-08T14:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.922119 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"390b82b2-610e-44c0-b509-c2e26db021e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb4658ef6d767e0b260c91435cffde1ce49a2ef282408f85c68c7b459323f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.937049 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.950907 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.982323 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38d73c12fc68f3781c7bc98848c8af66254da074464f5b735b44b3eb0afabdc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:01:47Z\\\",\\\"message\\\":\\\"cPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 14:01:47.550001 6445 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-operator for network=default are: map[]\\\\nF1008 14:01:47.549989 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:01:47Z is after 2025-08-24T17:21:41Z]\\\\nI1008 14:01:47.550011 6445 services_controller.go:443] Built service openshift-machine-config-operator/machine-config-operator LB cluster-wide configs for network=default: []services.lbConfig{se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:19Z\\\",\\\"message\\\":\\\"nNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:02:19.767861 6800 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:02:19.767873 6800 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 14:02:19.767890 6800 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 14:02:19.767902 6800 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 14:02:19.767911 6800 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 14:02:19.767927 6800 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:02:19.767944 6800 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 14:02:19.767957 6800 factory.go:656] Stopping watch factory\\\\nI1008 14:02:19.767964 6800 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 14:02:19.767978 6800 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 14:02:19.768004 6800 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 14:02:19.768012 6800 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 14:02:19.768342 6800 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:02:19.768387 6800 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:02:19.768524 6800 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:02:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:20 crc kubenswrapper[4789]: I1008 14:02:20.999528 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:20Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.019095 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.019123 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.019133 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.019150 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.019161 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.122746 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.122824 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.122845 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.122873 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.122891 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.225114 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.225160 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.225172 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.225192 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.225203 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.276984 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/3.log" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.280795 4789 scope.go:117] "RemoveContainer" containerID="3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298" Oct 08 14:02:21 crc kubenswrapper[4789]: E1008 14:02:21.281020 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.300914 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.316900 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.327407 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.327460 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.327483 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.327507 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.327523 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.329230 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.344196 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.360853 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.376810 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e0199c9-c939-49f7-991b-b9c6fef2f435\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.396818 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.415132 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.429458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.429506 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.429524 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.429547 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.429564 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.433107 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.448650 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.469471 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.483373 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:09Z\\\",\\\"message\\\":\\\"2025-10-08T14:01:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c\\\\n2025-10-08T14:01:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c to /host/opt/cni/bin/\\\\n2025-10-08T14:01:24Z [verbose] multus-daemon started\\\\n2025-10-08T14:01:24Z [verbose] Readiness Indicator file check\\\\n2025-10-08T14:02:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.495418 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"390b82b2-610e-44c0-b509-c2e26db021e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb4658ef6d767e0b260c91435cffde1ce49a2ef282408f85c68c7b459323f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.508627 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.521785 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.531962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.532214 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.532285 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.532449 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.532539 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.552366 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:19Z\\\",\\\"message\\\":\\\"nNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:02:19.767861 6800 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:02:19.767873 6800 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 14:02:19.767890 6800 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 14:02:19.767902 6800 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 14:02:19.767911 6800 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 14:02:19.767927 6800 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:02:19.767944 6800 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 14:02:19.767957 6800 factory.go:656] Stopping watch factory\\\\nI1008 14:02:19.767964 6800 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 14:02:19.767978 6800 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 14:02:19.768004 6800 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 14:02:19.768012 6800 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 14:02:19.768342 6800 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:02:19.768387 6800 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:02:19.768524 6800 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:02:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.568177 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:21Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.634561 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.634593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.634604 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.634620 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.634631 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.729087 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:21 crc kubenswrapper[4789]: E1008 14:02:21.729292 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.729087 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:21 crc kubenswrapper[4789]: E1008 14:02:21.729602 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.738012 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.738048 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.738059 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.738075 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.738087 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.840466 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.840575 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.840592 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.840617 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.840635 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.944535 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.944626 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.944644 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.944677 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:21 crc kubenswrapper[4789]: I1008 14:02:21.944715 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:21Z","lastTransitionTime":"2025-10-08T14:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.048694 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.048767 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.048789 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.048824 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.048848 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.151845 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.151897 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.151907 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.151925 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.151937 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.256674 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.256756 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.256782 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.256820 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.256848 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.360601 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.360682 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.360702 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.360734 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.360755 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.464051 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.464157 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.464179 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.464247 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.464269 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.568286 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.568336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.568350 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.568372 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.568386 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.672230 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.672292 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.672303 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.672328 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.672340 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.731308 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.731387 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:22 crc kubenswrapper[4789]: E1008 14:02:22.731558 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:22 crc kubenswrapper[4789]: E1008 14:02:22.731741 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.775784 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.775831 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.775842 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.775863 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.775875 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.879805 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.880194 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.880213 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.880242 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.880260 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.983742 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.983798 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.983817 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.983844 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:22 crc kubenswrapper[4789]: I1008 14:02:22.983864 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:22Z","lastTransitionTime":"2025-10-08T14:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.087384 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.087473 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.087497 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.087532 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.087558 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:23Z","lastTransitionTime":"2025-10-08T14:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.190945 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.191013 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.191027 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.191044 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.191062 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:23Z","lastTransitionTime":"2025-10-08T14:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.294228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.294346 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.294385 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.294426 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.294455 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:23Z","lastTransitionTime":"2025-10-08T14:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.398785 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.398880 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.398903 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.398932 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.398956 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:23Z","lastTransitionTime":"2025-10-08T14:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.501648 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.501718 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.501740 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.501769 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.501787 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:23Z","lastTransitionTime":"2025-10-08T14:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.604454 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.604516 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.604534 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.604564 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.604585 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:23Z","lastTransitionTime":"2025-10-08T14:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.708223 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.708295 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.708318 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.708351 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.708375 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:23Z","lastTransitionTime":"2025-10-08T14:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.729923 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.729971 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:23 crc kubenswrapper[4789]: E1008 14:02:23.730171 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:23 crc kubenswrapper[4789]: E1008 14:02:23.730329 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.811728 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.811796 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.811815 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.811842 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.811863 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:23Z","lastTransitionTime":"2025-10-08T14:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.915236 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.915294 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.915312 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.915328 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:23 crc kubenswrapper[4789]: I1008 14:02:23.915340 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:23Z","lastTransitionTime":"2025-10-08T14:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.018395 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.018477 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.018624 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.018653 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.018666 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.121626 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.121673 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.121682 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.121698 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.121709 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.224374 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.224424 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.224436 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.224452 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.224465 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.327092 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.327146 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.327159 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.327180 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.327193 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.430177 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.430252 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.430270 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.430301 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.430320 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.533125 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.533195 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.533220 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.533247 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.533266 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.636350 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.636402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.636419 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.636443 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.636462 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.728923 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.728970 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:24 crc kubenswrapper[4789]: E1008 14:02:24.729157 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:24 crc kubenswrapper[4789]: E1008 14:02:24.729262 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.739764 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.739815 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.739826 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.739841 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.739853 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.843295 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.843345 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.843361 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.843385 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.843402 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.946458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.946527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.946593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.946626 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:24 crc kubenswrapper[4789]: I1008 14:02:24.946652 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:24Z","lastTransitionTime":"2025-10-08T14:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.049326 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.049386 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.049404 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.049428 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.049446 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.152648 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.152762 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.152781 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.152806 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.152824 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.256029 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.256092 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.256111 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.256134 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.256151 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.359323 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.359391 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.359412 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.359442 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.359465 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.463078 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.463116 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.463125 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.463141 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.463152 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.565966 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.566016 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.566026 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.566159 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.566219 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.669777 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.669834 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.669851 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.669875 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.669892 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.684702 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.684856 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.684919 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.684965 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.685051 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685192 4789 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685200 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685217 4789 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685243 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685390 4789 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685185 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.685120856 +0000 UTC m=+149.591868398 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685541 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.685511546 +0000 UTC m=+149.592259298 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685614 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.685561207 +0000 UTC m=+149.592308969 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685641 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.685628149 +0000 UTC m=+149.592375681 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685251 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685712 4789 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685742 4789 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.685816 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.685801253 +0000 UTC m=+149.592548785 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.729845 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.729878 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.730155 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:25 crc kubenswrapper[4789]: E1008 14:02:25.730355 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.773405 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.773478 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.773490 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.773520 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.773535 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.876319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.876351 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.876359 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.876373 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.876384 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.978762 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.978822 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.978839 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.978863 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:25 crc kubenswrapper[4789]: I1008 14:02:25.978879 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:25Z","lastTransitionTime":"2025-10-08T14:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.082334 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.082439 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.082464 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.082495 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.082519 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:26Z","lastTransitionTime":"2025-10-08T14:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.184647 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.184700 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.184709 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.184723 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.184735 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:26Z","lastTransitionTime":"2025-10-08T14:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.288380 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.288423 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.288431 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.288444 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.288454 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:26Z","lastTransitionTime":"2025-10-08T14:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.390176 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.390214 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.390225 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.390240 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.390250 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:26Z","lastTransitionTime":"2025-10-08T14:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.492883 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.492922 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.492931 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.492962 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.493032 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:26Z","lastTransitionTime":"2025-10-08T14:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.597172 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.597228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.597242 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.597261 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.597276 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:26Z","lastTransitionTime":"2025-10-08T14:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.699722 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.699775 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.699790 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.699811 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.699825 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:26Z","lastTransitionTime":"2025-10-08T14:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.729309 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.729335 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:26 crc kubenswrapper[4789]: E1008 14:02:26.729501 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:26 crc kubenswrapper[4789]: E1008 14:02:26.729674 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.802104 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.802160 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.802177 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.802200 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.802218 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:26Z","lastTransitionTime":"2025-10-08T14:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.905226 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.905281 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.905298 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.905323 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:26 crc kubenswrapper[4789]: I1008 14:02:26.905343 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:26Z","lastTransitionTime":"2025-10-08T14:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.008212 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.008250 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.008294 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.008313 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.008324 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.111232 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.111289 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.111307 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.111334 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.111361 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.215572 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.215624 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.215636 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.215655 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.215667 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.318250 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.318297 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.318309 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.318327 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.318340 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.421605 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.421640 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.421650 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.421665 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.421675 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.529020 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.529085 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.529104 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.529130 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.529157 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.632085 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.632136 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.632152 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.632175 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.632194 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.729730 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.729753 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:27 crc kubenswrapper[4789]: E1008 14:02:27.730058 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:27 crc kubenswrapper[4789]: E1008 14:02:27.730198 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.734979 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.735026 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.735038 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.735054 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.735067 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.838103 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.838154 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.838173 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.838194 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.838214 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.941228 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.941271 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.941283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.941298 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:27 crc kubenswrapper[4789]: I1008 14:02:27.941310 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:27Z","lastTransitionTime":"2025-10-08T14:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.043916 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.044021 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.044048 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.044112 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.044135 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.146722 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.146772 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.146788 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.146805 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.146815 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.250457 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.250534 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.250558 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.250593 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.250621 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.353134 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.353190 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.353204 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.353220 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.353233 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.442977 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.443047 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.443063 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.443083 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.443096 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: E1008 14:02:28.463134 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.468267 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.468336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.468359 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.468386 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.468404 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: E1008 14:02:28.487686 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.492191 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.492244 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.492262 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.492284 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.492303 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: E1008 14:02:28.507652 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.512662 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.512732 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.512757 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.512790 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.512813 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: E1008 14:02:28.530716 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.535803 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.535856 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.535874 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.535898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.535916 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: E1008 14:02:28.556558 4789 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59598eb2-dd3c-404b-b870-d7b0c91c1ae8\\\",\\\"systemUUID\\\":\\\"3aa13740-a16f-4d58-8132-6848dda41690\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:28Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:28 crc kubenswrapper[4789]: E1008 14:02:28.556777 4789 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.558794 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.558852 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.558872 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.558897 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.558915 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.661421 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.661516 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.661534 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.661559 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.661589 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.729413 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.729489 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:28 crc kubenswrapper[4789]: E1008 14:02:28.729590 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:28 crc kubenswrapper[4789]: E1008 14:02:28.729728 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.764654 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.764763 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.764785 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.764807 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.764825 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.868828 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.868899 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.868916 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.868942 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.868959 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.972713 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.972782 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.972800 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.972828 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:28 crc kubenswrapper[4789]: I1008 14:02:28.972846 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:28Z","lastTransitionTime":"2025-10-08T14:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.076308 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.076379 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.076399 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.076429 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.076449 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:29Z","lastTransitionTime":"2025-10-08T14:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.179686 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.179757 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.179776 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.179801 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.179821 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:29Z","lastTransitionTime":"2025-10-08T14:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.282917 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.282968 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.283014 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.283035 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.283050 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:29Z","lastTransitionTime":"2025-10-08T14:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.386729 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.386795 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.386813 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.386836 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.386852 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:29Z","lastTransitionTime":"2025-10-08T14:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.489881 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.489920 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.489933 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.489950 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.489962 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:29Z","lastTransitionTime":"2025-10-08T14:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.592650 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.592717 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.592735 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.592760 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.592779 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:29Z","lastTransitionTime":"2025-10-08T14:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.695520 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.695590 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.695609 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.695632 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.695651 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:29Z","lastTransitionTime":"2025-10-08T14:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.729297 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:29 crc kubenswrapper[4789]: E1008 14:02:29.729475 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.729320 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:29 crc kubenswrapper[4789]: E1008 14:02:29.729701 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.798965 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.799068 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.799086 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.799108 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.799125 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:29Z","lastTransitionTime":"2025-10-08T14:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.901420 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.901459 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.901468 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.901482 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:29 crc kubenswrapper[4789]: I1008 14:02:29.901491 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:29Z","lastTransitionTime":"2025-10-08T14:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.004185 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.004279 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.004298 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.004321 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.004337 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.106810 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.106860 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.106872 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.106888 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.106900 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.208898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.208943 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.208984 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.209022 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.209034 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.312576 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.312613 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.312623 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.312637 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.312646 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.417041 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.417098 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.417113 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.417140 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.417155 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.520737 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.520818 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.520839 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.520868 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.520887 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.623218 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.623276 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.623286 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.623301 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.623312 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.727191 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.727250 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.727263 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.727283 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.727299 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.729726 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.729747 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:30 crc kubenswrapper[4789]: E1008 14:02:30.729940 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:30 crc kubenswrapper[4789]: E1008 14:02:30.730055 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.751532 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.780251 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bee958d-f383-4e01-8379-b3d2ba664129\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:19Z\\\",\\\"message\\\":\\\"nNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 14:02:19.767861 6800 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1008 14:02:19.767873 6800 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 14:02:19.767890 6800 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 14:02:19.767902 6800 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 14:02:19.767911 6800 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 14:02:19.767927 6800 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 14:02:19.767944 6800 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 14:02:19.767957 6800 factory.go:656] Stopping watch factory\\\\nI1008 14:02:19.767964 6800 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 14:02:19.767978 6800 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 14:02:19.768004 6800 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 14:02:19.768012 6800 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 14:02:19.768342 6800 ovnkube.go:599] Stopped ovnkube\\\\nI1008 14:02:19.768387 6800 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 14:02:19.768524 6800 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:02:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5zkqv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-9gpb2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.797610 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tf2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c453426b-c6f6-4d8f-a301-5b70c2c183a1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://543ee1eb0617f6c95c43a3acd3833af7cffe418d535c299283efab47b0d15eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zcg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tf2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.817456 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"390b82b2-610e-44c0-b509-c2e26db021e1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcb4658ef6d767e0b260c91435cffde1ce49a2ef282408f85c68c7b459323f4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09d982d82eaf550b89cf9d7b9a51fe3aa87ed0e9c78de53ca8ed23678014fdd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.830039 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.830105 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.830126 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.830152 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.830175 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.841855 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6fdaa9a95b6bc04093c278a3924d81d71eda5ac28d971e9f9e13075c6f71379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.855181 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b269s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"765705a4-a303-4281-9677-5d0769f6d157\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t54j9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b269s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.870968 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0b2f620-4874-453a-a61a-c113800bcc82\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://326ac9d7d420b0aa02d77d0a0570f879bf37be2ad129e11d8c7dfc9472ea84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://566c24368421cb55717cadb106f2a576661b0ab4e48599aedad41cd3c41ca402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75f9e916590f35fa53b25550b50882d211a27c3db1f872568a3230a74dc64613\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://46a452313fdb4a4f369d63c34bec91968a32dda82b70606477cfce1e6e42e19d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099a0213755197b5f121430603206b38d4a51363e3bf055525fa24040ffcc221\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T14:01:20Z\\\",\\\"message\\\":\\\"08 14:01:20.001955 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 14:01:20.001960 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 14:01:20.001964 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 14:01:20.002225 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1008 14:01:20.006761 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759932063\\\\\\\\\\\\\\\" (2025-10-08 14:01:03 +0000 UTC to 2025-11-07 14:01:04 +0000 UTC (now=2025-10-08 14:01:20.006725042 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006927 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759932074\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759932074\\\\\\\\\\\\\\\" (2025-10-08 13:01:14 +0000 UTC to 2026-10-08 13:01:14 +0000 UTC (now=2025-10-08 14:01:20.006904847 +0000 UTC))\\\\\\\"\\\\nI1008 14:01:20.006950 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1008 14:01:20.006974 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1008 14:01:20.007019 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007044 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1008 14:01:20.007090 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4010190347/tls.crt::/tmp/serving-cert-4010190347/tls.key\\\\\\\"\\\\nI1008 14:01:20.007212 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1008 14:01:20.009056 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b52ba7e4f83b2438a2f9184e929ea9288d0233b9a2034db48354b6634111d20\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7672253ac15b364e55f637f0a7221c5311927a98a292340d15adcf260a4bbf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.890214 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.905321 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95e63cbd-0d26-4d8e-b6e2-768355e7dcca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2378addb90868162487a23697cf9c28b51cf6a41d0fc4d59ffadc20989a21bf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ec33d14ce26849c15f760ef5d24f46445c6d61ac14e570e7c1c99798204ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqgdp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xhhmr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.922765 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b2e313a-834b-47fb-adae-992853bf6a6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa3709e73cc186717acb5925d843668aa8f97d9baf3755c1bbffba844a94f9c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tb5jh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-94gx2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.933161 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.933210 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.933220 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.933240 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.933252 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:30Z","lastTransitionTime":"2025-10-08T14:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.939240 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.952500 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r858x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de6093ed-9813-429d-812d-26bddadd2fa6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23d6e0a12eadfc4a7221333608eef5a48ea5893e7babcffa7f12624e32abd9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5bxb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r858x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.970077 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d006f91-6689-46a3-a6cd-7ed9f2fd9188\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb18f09789a305fbc6649afd6cf5fd0f275ec2d63928dfae5bcbadbd7e4c8658\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://37affd74aac6a8d398c3ccb6c0a89a7ec70168790db5bec7246592eb0802d7e4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://04f2bdf1ceccfe7690d895005c5834341558978afb2bd08b0863f06eb6f0e121\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12de12f2ed2e06ba79a0a1057e1a5372f01492b54b954f25a82c24ca588e853c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5df4aa08cb4283fb920de24523fbf47699e19d34d6d32bd047f727b2e33f684\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3bcba884cc493bf77907178c6c663bcaf2a540b3445a590af61a5e892144e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50a6bc0154cc81201ad1c35c53c30a027648dfcb829922b53df61a20fa941431\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cksqm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4m4jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:30 crc kubenswrapper[4789]: I1008 14:02:30.992661 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5428v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4aee6ee4-49ed-4135-8bb0-003eb4f70f05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:02:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T14:02:09Z\\\",\\\"message\\\":\\\"2025-10-08T14:01:23+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c\\\\n2025-10-08T14:01:23+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_4b2c3a98-ca7e-4df5-8bb0-c9052e0adb4c to /host/opt/cni/bin/\\\\n2025-10-08T14:01:24Z [verbose] multus-daemon started\\\\n2025-10-08T14:01:24Z [verbose] Readiness Indicator file check\\\\n2025-10-08T14:02:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:02:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s4w64\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:21Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5428v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:30Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.007752 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6e0199c9-c939-49f7-991b-b9c6fef2f435\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cbf3ae32dd49bbd0ee68d33a8b79a881d0bf3a989acd9004dac0ce134bf499a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d51bd79556f4984768a176a2e25e76c7dfa440567c0f6b26c39129a7ac133fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9057d1f9b6ef37e4d5b495f4b2eb4195a3e2a06aa5f8a5292ec39921eea6ce82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a61d66e6a9ece66998f39b2968b915752da1e9cbd719370e88dd205f95249df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T14:01:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T14:01:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T14:01:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.027580 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71755801dc6d8e1bb54209b51375bd0706452dd2c9c8d8b07da3af9f9e5bcf58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.035461 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.035524 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.035539 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.035562 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.035579 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.042349 4789 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T14:01:22Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb4ffcdc695a8aa5ac2cc21c542c41e6bdc50049dd8748f9f74178a0df0eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e482e8d7c0e92cb124872efa5f427a56a011e495afc05f808b08381fccfa883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T14:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T14:02:31Z is after 2025-08-24T17:21:41Z" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.138839 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.138889 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.138897 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.138913 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.138922 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.240894 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.240938 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.240950 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.240975 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.241008 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.343656 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.343697 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.343710 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.343730 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.343743 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.446924 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.446968 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.446979 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.447026 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.447039 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.550475 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.550539 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.550556 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.550583 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.550600 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.658105 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.658181 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.658199 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.658227 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.658244 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.730210 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.730249 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:31 crc kubenswrapper[4789]: E1008 14:02:31.730404 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:31 crc kubenswrapper[4789]: E1008 14:02:31.730629 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.761733 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.761813 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.761827 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.761846 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.761859 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.865810 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.866352 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.866416 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.866459 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.866485 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.970377 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.970441 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.970458 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.970483 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:31 crc kubenswrapper[4789]: I1008 14:02:31.970504 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:31Z","lastTransitionTime":"2025-10-08T14:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.073827 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.073896 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.073919 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.073947 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.073965 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:32Z","lastTransitionTime":"2025-10-08T14:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.176724 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.176771 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.176780 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.176796 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.176807 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:32Z","lastTransitionTime":"2025-10-08T14:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.280116 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.280212 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.280238 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.280276 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.280301 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:32Z","lastTransitionTime":"2025-10-08T14:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.384446 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.384509 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.384533 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.384562 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.384584 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:32Z","lastTransitionTime":"2025-10-08T14:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.488536 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.488614 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.488649 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.488683 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.488712 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:32Z","lastTransitionTime":"2025-10-08T14:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.592257 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.592319 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.592331 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.592348 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.592360 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:32Z","lastTransitionTime":"2025-10-08T14:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.696089 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.696150 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.696389 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.697530 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.697558 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:32Z","lastTransitionTime":"2025-10-08T14:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.730078 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:32 crc kubenswrapper[4789]: E1008 14:02:32.730307 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.730340 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:32 crc kubenswrapper[4789]: E1008 14:02:32.730692 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.731946 4789 scope.go:117] "RemoveContainer" containerID="3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298" Oct 08 14:02:32 crc kubenswrapper[4789]: E1008 14:02:32.732972 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.800629 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.800693 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.800712 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.800735 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.800755 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:32Z","lastTransitionTime":"2025-10-08T14:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.907917 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.907981 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.908014 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.908032 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:32 crc kubenswrapper[4789]: I1008 14:02:32.908045 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:32Z","lastTransitionTime":"2025-10-08T14:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.011527 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.011589 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.011605 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.011629 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.011647 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.115393 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.115493 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.115513 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.115538 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.115556 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.218775 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.218847 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.218865 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.218893 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.218913 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.323201 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.323274 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.323304 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.323339 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.323363 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.427296 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.427346 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.427356 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.427373 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.427384 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.530417 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.530464 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.530477 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.530496 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.530506 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.637712 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.638207 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.638234 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.638342 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.638413 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.729216 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.729238 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:33 crc kubenswrapper[4789]: E1008 14:02:33.729452 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:33 crc kubenswrapper[4789]: E1008 14:02:33.729652 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.741406 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.741494 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.741523 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.741554 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.741573 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.844202 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.844245 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.844254 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.844270 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.844282 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.947119 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.947163 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.947174 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.947190 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:33 crc kubenswrapper[4789]: I1008 14:02:33.947202 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:33Z","lastTransitionTime":"2025-10-08T14:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.050371 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.050430 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.050439 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.050456 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.050467 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.153478 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.153563 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.153583 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.153612 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.153634 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.256898 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.256935 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.256943 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.256957 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.256966 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.360118 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.360172 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.360185 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.360205 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.360223 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.463161 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.463249 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.463270 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.463309 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.463334 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.566044 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.566105 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.566122 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.566149 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.566166 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.669856 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.669902 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.669911 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.669930 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.669941 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.730032 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.730172 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:34 crc kubenswrapper[4789]: E1008 14:02:34.730326 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:34 crc kubenswrapper[4789]: E1008 14:02:34.730514 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.773860 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.773967 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.773981 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.774047 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.774059 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.875914 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.875981 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.876031 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.876057 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.876073 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.979099 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.979164 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.979181 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.979205 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:34 crc kubenswrapper[4789]: I1008 14:02:34.979222 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:34Z","lastTransitionTime":"2025-10-08T14:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.081752 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.081804 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.081825 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.081852 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.081872 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:35Z","lastTransitionTime":"2025-10-08T14:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.184949 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.185065 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.185088 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.185121 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.185146 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:35Z","lastTransitionTime":"2025-10-08T14:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.288039 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.288107 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.288123 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.288150 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.288168 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:35Z","lastTransitionTime":"2025-10-08T14:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.390763 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.390819 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.390832 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.390855 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.390868 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:35Z","lastTransitionTime":"2025-10-08T14:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.493146 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.493210 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.493227 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.493252 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.493269 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:35Z","lastTransitionTime":"2025-10-08T14:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.596673 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.596731 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.596741 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.596756 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.596765 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:35Z","lastTransitionTime":"2025-10-08T14:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.698635 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.698696 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.698706 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.698720 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.698733 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:35Z","lastTransitionTime":"2025-10-08T14:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.729926 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.729928 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:35 crc kubenswrapper[4789]: E1008 14:02:35.730053 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:35 crc kubenswrapper[4789]: E1008 14:02:35.730254 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.801720 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.801976 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.802008 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.802028 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.802042 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:35Z","lastTransitionTime":"2025-10-08T14:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.904536 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.904583 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.904595 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.904612 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:35 crc kubenswrapper[4789]: I1008 14:02:35.904624 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:35Z","lastTransitionTime":"2025-10-08T14:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.007758 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.007832 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.007851 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.007874 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.007890 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.110029 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.110073 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.110081 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.110096 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.110104 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.212897 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.212964 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.212979 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.213038 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.213052 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.320090 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.320351 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.320450 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.320596 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.320698 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.423956 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.424299 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.424384 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.424485 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.424560 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.528235 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.528323 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.528336 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.528359 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.528372 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.632048 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.632133 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.632156 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.632192 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.632218 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.730020 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.730056 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:36 crc kubenswrapper[4789]: E1008 14:02:36.730263 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:36 crc kubenswrapper[4789]: E1008 14:02:36.730438 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.734497 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.734562 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.734581 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.734607 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.734626 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.752371 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.838161 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.838197 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.838212 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.838226 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.838235 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.941524 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.941623 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.941642 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.941669 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:36 crc kubenswrapper[4789]: I1008 14:02:36.941686 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:36Z","lastTransitionTime":"2025-10-08T14:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.044371 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.044448 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.044468 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.044503 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.044521 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.147334 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.147386 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.147402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.147427 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.147441 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.251055 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.251118 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.251131 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.251152 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.251169 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.353346 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.353402 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.353417 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.353438 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.353452 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.456444 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.456506 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.456518 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.456541 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.456553 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.559653 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.559763 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.559785 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.559822 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.559844 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.662851 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.662915 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.662927 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.662952 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.662964 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.728906 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:37 crc kubenswrapper[4789]: E1008 14:02:37.729134 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.729241 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:37 crc kubenswrapper[4789]: E1008 14:02:37.729687 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.765633 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.765713 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.765735 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.765767 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.765787 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.869538 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.869617 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.869632 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.869659 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.869675 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.972062 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.972106 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.972118 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.972135 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:37 crc kubenswrapper[4789]: I1008 14:02:37.972146 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:37Z","lastTransitionTime":"2025-10-08T14:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.076722 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.076824 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.076849 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.076888 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.076914 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:38Z","lastTransitionTime":"2025-10-08T14:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.180032 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.180105 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.180161 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.180195 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.180218 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:38Z","lastTransitionTime":"2025-10-08T14:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.284160 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.284225 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.284242 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.284268 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.284287 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:38Z","lastTransitionTime":"2025-10-08T14:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.386974 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.387070 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.387105 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.387128 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.387145 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:38Z","lastTransitionTime":"2025-10-08T14:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.490081 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.490155 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.490178 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.490212 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.490237 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:38Z","lastTransitionTime":"2025-10-08T14:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.595474 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.595567 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.595589 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.595627 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.595659 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:38Z","lastTransitionTime":"2025-10-08T14:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.653829 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.653894 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.653913 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.653960 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.654048 4789 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T14:02:38Z","lastTransitionTime":"2025-10-08T14:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.725460 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs"] Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.725943 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.729355 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:38 crc kubenswrapper[4789]: E1008 14:02:38.729583 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.729654 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.729891 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.729946 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 08 14:02:38 crc kubenswrapper[4789]: E1008 14:02:38.730037 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.732543 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.732692 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.736730 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/92dbcf36-2c5a-42ce-bb9f-0783757e7060-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.736770 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92dbcf36-2c5a-42ce-bb9f-0783757e7060-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.736803 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/92dbcf36-2c5a-42ce-bb9f-0783757e7060-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.736846 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/92dbcf36-2c5a-42ce-bb9f-0783757e7060-service-ca\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.736924 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/92dbcf36-2c5a-42ce-bb9f-0783757e7060-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.794635 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tf2m8" podStartSLOduration=78.794613814 podStartE2EDuration="1m18.794613814s" podCreationTimestamp="2025-10-08 14:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:38.794292036 +0000 UTC m=+98.701039538" watchObservedRunningTime="2025-10-08 14:02:38.794613814 +0000 UTC m=+98.701361306" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.804484 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=28.804469413 podStartE2EDuration="28.804469413s" podCreationTimestamp="2025-10-08 14:02:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:38.80353528 +0000 UTC m=+98.710282772" watchObservedRunningTime="2025-10-08 14:02:38.804469413 +0000 UTC m=+98.711216905" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.837677 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/92dbcf36-2c5a-42ce-bb9f-0783757e7060-service-ca\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.837760 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/92dbcf36-2c5a-42ce-bb9f-0783757e7060-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.837811 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/92dbcf36-2c5a-42ce-bb9f-0783757e7060-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.837855 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/92dbcf36-2c5a-42ce-bb9f-0783757e7060-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.837880 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92dbcf36-2c5a-42ce-bb9f-0783757e7060-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.837911 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:38 crc kubenswrapper[4789]: E1008 14:02:38.838083 4789 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:02:38 crc kubenswrapper[4789]: E1008 14:02:38.838150 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs podName:765705a4-a303-4281-9677-5d0769f6d157 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:42.838134043 +0000 UTC m=+162.744881535 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs") pod "network-metrics-daemon-b269s" (UID: "765705a4-a303-4281-9677-5d0769f6d157") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.839148 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/92dbcf36-2c5a-42ce-bb9f-0783757e7060-service-ca\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.839202 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/92dbcf36-2c5a-42ce-bb9f-0783757e7060-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.839489 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/92dbcf36-2c5a-42ce-bb9f-0783757e7060-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.855274 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92dbcf36-2c5a-42ce-bb9f-0783757e7060-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.864411 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/92dbcf36-2c5a-42ce-bb9f-0783757e7060-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-945vs\" (UID: \"92dbcf36-2c5a-42ce-bb9f-0783757e7060\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.871338 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.871305011 podStartE2EDuration="1m17.871305011s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:38.85821989 +0000 UTC m=+98.764967392" watchObservedRunningTime="2025-10-08 14:02:38.871305011 +0000 UTC m=+98.778052563" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.952130 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xhhmr" podStartSLOduration=77.952109541 podStartE2EDuration="1m17.952109541s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:38.891947382 +0000 UTC m=+98.798694874" watchObservedRunningTime="2025-10-08 14:02:38.952109541 +0000 UTC m=+98.858857033" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.952232 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=2.952229784 podStartE2EDuration="2.952229784s" podCreationTimestamp="2025-10-08 14:02:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:38.951396213 +0000 UTC m=+98.858143695" watchObservedRunningTime="2025-10-08 14:02:38.952229784 +0000 UTC m=+98.858977276" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.966632 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podStartSLOduration=77.966609067 podStartE2EDuration="1m17.966609067s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:38.966561486 +0000 UTC m=+98.873308978" watchObservedRunningTime="2025-10-08 14:02:38.966609067 +0000 UTC m=+98.873356549" Oct 08 14:02:38 crc kubenswrapper[4789]: I1008 14:02:38.991698 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-r858x" podStartSLOduration=79.99168013 podStartE2EDuration="1m19.99168013s" podCreationTimestamp="2025-10-08 14:01:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:38.991513986 +0000 UTC m=+98.898261478" watchObservedRunningTime="2025-10-08 14:02:38.99168013 +0000 UTC m=+98.898427622" Oct 08 14:02:39 crc kubenswrapper[4789]: I1008 14:02:39.007138 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-4m4jj" podStartSLOduration=78.00712173 podStartE2EDuration="1m18.00712173s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:39.006159746 +0000 UTC m=+98.912907248" watchObservedRunningTime="2025-10-08 14:02:39.00712173 +0000 UTC m=+98.913869222" Oct 08 14:02:39 crc kubenswrapper[4789]: I1008 14:02:39.022354 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-5428v" podStartSLOduration=78.022329464 podStartE2EDuration="1m18.022329464s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:39.022060677 +0000 UTC m=+98.928808169" watchObservedRunningTime="2025-10-08 14:02:39.022329464 +0000 UTC m=+98.929076956" Oct 08 14:02:39 crc kubenswrapper[4789]: I1008 14:02:39.042519 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" Oct 08 14:02:39 crc kubenswrapper[4789]: I1008 14:02:39.062323 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=43.062296363 podStartE2EDuration="43.062296363s" podCreationTimestamp="2025-10-08 14:01:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:39.06020833 +0000 UTC m=+98.966955852" watchObservedRunningTime="2025-10-08 14:02:39.062296363 +0000 UTC m=+98.969043885" Oct 08 14:02:39 crc kubenswrapper[4789]: I1008 14:02:39.344525 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" event={"ID":"92dbcf36-2c5a-42ce-bb9f-0783757e7060","Type":"ContainerStarted","Data":"ccc090ea8a71a7b1f98b03c2bd1341ba3261fc770f5ca4b8917de6515430e7d1"} Oct 08 14:02:39 crc kubenswrapper[4789]: I1008 14:02:39.344580 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" event={"ID":"92dbcf36-2c5a-42ce-bb9f-0783757e7060","Type":"ContainerStarted","Data":"be118e9836280466428681b99a8ed36e321d5e1f5db595f6c6f7aa06ea0756eb"} Oct 08 14:02:39 crc kubenswrapper[4789]: I1008 14:02:39.375970 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-945vs" podStartSLOduration=78.375937292 podStartE2EDuration="1m18.375937292s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:39.375496921 +0000 UTC m=+99.282244413" watchObservedRunningTime="2025-10-08 14:02:39.375937292 +0000 UTC m=+99.282684814" Oct 08 14:02:39 crc kubenswrapper[4789]: I1008 14:02:39.729291 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:39 crc kubenswrapper[4789]: I1008 14:02:39.729299 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:39 crc kubenswrapper[4789]: E1008 14:02:39.729598 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:39 crc kubenswrapper[4789]: E1008 14:02:39.729676 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:40 crc kubenswrapper[4789]: I1008 14:02:40.729583 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:40 crc kubenswrapper[4789]: E1008 14:02:40.732068 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:40 crc kubenswrapper[4789]: I1008 14:02:40.732140 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:40 crc kubenswrapper[4789]: E1008 14:02:40.732344 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:41 crc kubenswrapper[4789]: I1008 14:02:41.729534 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:41 crc kubenswrapper[4789]: I1008 14:02:41.729620 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:41 crc kubenswrapper[4789]: E1008 14:02:41.729684 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:41 crc kubenswrapper[4789]: E1008 14:02:41.729765 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:42 crc kubenswrapper[4789]: I1008 14:02:42.729215 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:42 crc kubenswrapper[4789]: E1008 14:02:42.729352 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:42 crc kubenswrapper[4789]: I1008 14:02:42.729409 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:42 crc kubenswrapper[4789]: E1008 14:02:42.729633 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:43 crc kubenswrapper[4789]: I1008 14:02:43.730018 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:43 crc kubenswrapper[4789]: I1008 14:02:43.730041 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:43 crc kubenswrapper[4789]: E1008 14:02:43.730190 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:43 crc kubenswrapper[4789]: E1008 14:02:43.730634 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:43 crc kubenswrapper[4789]: I1008 14:02:43.730823 4789 scope.go:117] "RemoveContainer" containerID="3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298" Oct 08 14:02:43 crc kubenswrapper[4789]: E1008 14:02:43.730976 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:02:44 crc kubenswrapper[4789]: I1008 14:02:44.729633 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:44 crc kubenswrapper[4789]: I1008 14:02:44.729639 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:44 crc kubenswrapper[4789]: E1008 14:02:44.730151 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:44 crc kubenswrapper[4789]: E1008 14:02:44.730326 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:45 crc kubenswrapper[4789]: I1008 14:02:45.729492 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:45 crc kubenswrapper[4789]: I1008 14:02:45.729573 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:45 crc kubenswrapper[4789]: E1008 14:02:45.729654 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:45 crc kubenswrapper[4789]: E1008 14:02:45.729777 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:46 crc kubenswrapper[4789]: I1008 14:02:46.729898 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:46 crc kubenswrapper[4789]: I1008 14:02:46.731061 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:46 crc kubenswrapper[4789]: E1008 14:02:46.731096 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:46 crc kubenswrapper[4789]: E1008 14:02:46.731333 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:47 crc kubenswrapper[4789]: I1008 14:02:47.729226 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:47 crc kubenswrapper[4789]: E1008 14:02:47.729437 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:47 crc kubenswrapper[4789]: I1008 14:02:47.729779 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:47 crc kubenswrapper[4789]: E1008 14:02:47.729875 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:48 crc kubenswrapper[4789]: I1008 14:02:48.729509 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:48 crc kubenswrapper[4789]: I1008 14:02:48.729608 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:48 crc kubenswrapper[4789]: E1008 14:02:48.730461 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:48 crc kubenswrapper[4789]: E1008 14:02:48.730585 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:49 crc kubenswrapper[4789]: I1008 14:02:49.729698 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:49 crc kubenswrapper[4789]: I1008 14:02:49.729720 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:49 crc kubenswrapper[4789]: E1008 14:02:49.730115 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:49 crc kubenswrapper[4789]: E1008 14:02:49.730368 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:49 crc kubenswrapper[4789]: I1008 14:02:49.749143 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 08 14:02:50 crc kubenswrapper[4789]: I1008 14:02:50.729181 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:50 crc kubenswrapper[4789]: I1008 14:02:50.729206 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:50 crc kubenswrapper[4789]: E1008 14:02:50.731032 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:50 crc kubenswrapper[4789]: E1008 14:02:50.731147 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:50 crc kubenswrapper[4789]: I1008 14:02:50.754629 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=1.754599298 podStartE2EDuration="1.754599298s" podCreationTimestamp="2025-10-08 14:02:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:02:50.754545526 +0000 UTC m=+110.661293018" watchObservedRunningTime="2025-10-08 14:02:50.754599298 +0000 UTC m=+110.661346800" Oct 08 14:02:51 crc kubenswrapper[4789]: I1008 14:02:51.729084 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:51 crc kubenswrapper[4789]: I1008 14:02:51.729084 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:51 crc kubenswrapper[4789]: E1008 14:02:51.729312 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:51 crc kubenswrapper[4789]: E1008 14:02:51.729375 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:52 crc kubenswrapper[4789]: I1008 14:02:52.729523 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:52 crc kubenswrapper[4789]: I1008 14:02:52.729729 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:52 crc kubenswrapper[4789]: E1008 14:02:52.729810 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:52 crc kubenswrapper[4789]: E1008 14:02:52.729717 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:53 crc kubenswrapper[4789]: I1008 14:02:53.729754 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:53 crc kubenswrapper[4789]: I1008 14:02:53.729870 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:53 crc kubenswrapper[4789]: E1008 14:02:53.729957 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:53 crc kubenswrapper[4789]: E1008 14:02:53.730107 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:54 crc kubenswrapper[4789]: I1008 14:02:54.729860 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:54 crc kubenswrapper[4789]: I1008 14:02:54.729943 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:54 crc kubenswrapper[4789]: E1008 14:02:54.730079 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:54 crc kubenswrapper[4789]: E1008 14:02:54.730171 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:55 crc kubenswrapper[4789]: I1008 14:02:55.729500 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:55 crc kubenswrapper[4789]: I1008 14:02:55.729651 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:55 crc kubenswrapper[4789]: E1008 14:02:55.729740 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:55 crc kubenswrapper[4789]: E1008 14:02:55.729843 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:56 crc kubenswrapper[4789]: I1008 14:02:56.405976 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/1.log" Oct 08 14:02:56 crc kubenswrapper[4789]: I1008 14:02:56.406388 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/0.log" Oct 08 14:02:56 crc kubenswrapper[4789]: I1008 14:02:56.406423 4789 generic.go:334] "Generic (PLEG): container finished" podID="4aee6ee4-49ed-4135-8bb0-003eb4f70f05" containerID="25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf" exitCode=1 Oct 08 14:02:56 crc kubenswrapper[4789]: I1008 14:02:56.406457 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5428v" event={"ID":"4aee6ee4-49ed-4135-8bb0-003eb4f70f05","Type":"ContainerDied","Data":"25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf"} Oct 08 14:02:56 crc kubenswrapper[4789]: I1008 14:02:56.406495 4789 scope.go:117] "RemoveContainer" containerID="b83ab26251363b3154dca6238a986f84b0459a005f78cafdd450ad8b47649b29" Oct 08 14:02:56 crc kubenswrapper[4789]: I1008 14:02:56.407141 4789 scope.go:117] "RemoveContainer" containerID="25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf" Oct 08 14:02:56 crc kubenswrapper[4789]: E1008 14:02:56.407473 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-5428v_openshift-multus(4aee6ee4-49ed-4135-8bb0-003eb4f70f05)\"" pod="openshift-multus/multus-5428v" podUID="4aee6ee4-49ed-4135-8bb0-003eb4f70f05" Oct 08 14:02:56 crc kubenswrapper[4789]: I1008 14:02:56.729783 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:56 crc kubenswrapper[4789]: I1008 14:02:56.729909 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:56 crc kubenswrapper[4789]: E1008 14:02:56.730095 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:56 crc kubenswrapper[4789]: E1008 14:02:56.730389 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:57 crc kubenswrapper[4789]: I1008 14:02:57.410386 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/1.log" Oct 08 14:02:57 crc kubenswrapper[4789]: I1008 14:02:57.729414 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:57 crc kubenswrapper[4789]: I1008 14:02:57.729469 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:57 crc kubenswrapper[4789]: E1008 14:02:57.729564 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:57 crc kubenswrapper[4789]: E1008 14:02:57.730194 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:02:57 crc kubenswrapper[4789]: I1008 14:02:57.730818 4789 scope.go:117] "RemoveContainer" containerID="3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298" Oct 08 14:02:57 crc kubenswrapper[4789]: E1008 14:02:57.731129 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-9gpb2_openshift-ovn-kubernetes(1bee958d-f383-4e01-8379-b3d2ba664129)\"" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" Oct 08 14:02:58 crc kubenswrapper[4789]: I1008 14:02:58.729916 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:02:58 crc kubenswrapper[4789]: I1008 14:02:58.729970 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:02:58 crc kubenswrapper[4789]: E1008 14:02:58.730101 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:02:58 crc kubenswrapper[4789]: E1008 14:02:58.730295 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:02:59 crc kubenswrapper[4789]: I1008 14:02:59.728970 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:02:59 crc kubenswrapper[4789]: E1008 14:02:59.729412 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:02:59 crc kubenswrapper[4789]: I1008 14:02:59.729099 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:02:59 crc kubenswrapper[4789]: E1008 14:02:59.729632 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:03:00 crc kubenswrapper[4789]: I1008 14:03:00.729488 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:00 crc kubenswrapper[4789]: E1008 14:03:00.730454 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:03:00 crc kubenswrapper[4789]: I1008 14:03:00.730503 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:00 crc kubenswrapper[4789]: E1008 14:03:00.730670 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:03:00 crc kubenswrapper[4789]: E1008 14:03:00.762345 4789 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 08 14:03:00 crc kubenswrapper[4789]: E1008 14:03:00.833716 4789 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 14:03:01 crc kubenswrapper[4789]: I1008 14:03:01.729803 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:01 crc kubenswrapper[4789]: I1008 14:03:01.729942 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:01 crc kubenswrapper[4789]: E1008 14:03:01.730241 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:03:01 crc kubenswrapper[4789]: E1008 14:03:01.730020 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:03:02 crc kubenswrapper[4789]: I1008 14:03:02.729242 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:02 crc kubenswrapper[4789]: I1008 14:03:02.729325 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:02 crc kubenswrapper[4789]: E1008 14:03:02.729410 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:03:02 crc kubenswrapper[4789]: E1008 14:03:02.729483 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:03:03 crc kubenswrapper[4789]: I1008 14:03:03.729793 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:03 crc kubenswrapper[4789]: I1008 14:03:03.729831 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:03 crc kubenswrapper[4789]: E1008 14:03:03.729957 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:03:03 crc kubenswrapper[4789]: E1008 14:03:03.730083 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:03:04 crc kubenswrapper[4789]: I1008 14:03:04.729767 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:04 crc kubenswrapper[4789]: I1008 14:03:04.729866 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:04 crc kubenswrapper[4789]: E1008 14:03:04.729905 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:03:04 crc kubenswrapper[4789]: E1008 14:03:04.730094 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:03:05 crc kubenswrapper[4789]: I1008 14:03:05.729432 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:05 crc kubenswrapper[4789]: I1008 14:03:05.729479 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:05 crc kubenswrapper[4789]: E1008 14:03:05.730020 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:03:05 crc kubenswrapper[4789]: E1008 14:03:05.730019 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:03:05 crc kubenswrapper[4789]: E1008 14:03:05.835051 4789 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 14:03:06 crc kubenswrapper[4789]: I1008 14:03:06.729018 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:06 crc kubenswrapper[4789]: I1008 14:03:06.729129 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:06 crc kubenswrapper[4789]: E1008 14:03:06.729563 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:03:06 crc kubenswrapper[4789]: E1008 14:03:06.729608 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:03:07 crc kubenswrapper[4789]: I1008 14:03:07.729320 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:07 crc kubenswrapper[4789]: E1008 14:03:07.729551 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:03:07 crc kubenswrapper[4789]: I1008 14:03:07.729341 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:07 crc kubenswrapper[4789]: I1008 14:03:07.729721 4789 scope.go:117] "RemoveContainer" containerID="25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf" Oct 08 14:03:07 crc kubenswrapper[4789]: E1008 14:03:07.729732 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:03:08 crc kubenswrapper[4789]: I1008 14:03:08.451597 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/1.log" Oct 08 14:03:08 crc kubenswrapper[4789]: I1008 14:03:08.452032 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5428v" event={"ID":"4aee6ee4-49ed-4135-8bb0-003eb4f70f05","Type":"ContainerStarted","Data":"b042dde4e6bd1efe56d4753229b6c7f56e3405e03b26aa92a893ee000bf1a70e"} Oct 08 14:03:08 crc kubenswrapper[4789]: I1008 14:03:08.729966 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:08 crc kubenswrapper[4789]: I1008 14:03:08.729968 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:08 crc kubenswrapper[4789]: E1008 14:03:08.730183 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:03:08 crc kubenswrapper[4789]: E1008 14:03:08.730291 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:03:09 crc kubenswrapper[4789]: I1008 14:03:09.729804 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:09 crc kubenswrapper[4789]: I1008 14:03:09.729804 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:09 crc kubenswrapper[4789]: E1008 14:03:09.730116 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:03:09 crc kubenswrapper[4789]: E1008 14:03:09.730255 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:03:10 crc kubenswrapper[4789]: I1008 14:03:10.729302 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:10 crc kubenswrapper[4789]: I1008 14:03:10.729301 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:10 crc kubenswrapper[4789]: E1008 14:03:10.731423 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:03:10 crc kubenswrapper[4789]: E1008 14:03:10.732193 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:03:10 crc kubenswrapper[4789]: I1008 14:03:10.733154 4789 scope.go:117] "RemoveContainer" containerID="3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298" Oct 08 14:03:10 crc kubenswrapper[4789]: E1008 14:03:10.836571 4789 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 14:03:11 crc kubenswrapper[4789]: I1008 14:03:11.466489 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/3.log" Oct 08 14:03:11 crc kubenswrapper[4789]: I1008 14:03:11.474149 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerStarted","Data":"135f63ea6d151315e8340b799662ffcfc90c6d9ecac7d612944bf3a97f65a626"} Oct 08 14:03:11 crc kubenswrapper[4789]: I1008 14:03:11.474595 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:03:11 crc kubenswrapper[4789]: I1008 14:03:11.729688 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:11 crc kubenswrapper[4789]: E1008 14:03:11.729855 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:03:11 crc kubenswrapper[4789]: I1008 14:03:11.730064 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:11 crc kubenswrapper[4789]: E1008 14:03:11.730142 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:03:11 crc kubenswrapper[4789]: I1008 14:03:11.732404 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podStartSLOduration=110.732389404 podStartE2EDuration="1m50.732389404s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:11.507644509 +0000 UTC m=+131.414392011" watchObservedRunningTime="2025-10-08 14:03:11.732389404 +0000 UTC m=+131.639136896" Oct 08 14:03:11 crc kubenswrapper[4789]: I1008 14:03:11.733314 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b269s"] Oct 08 14:03:11 crc kubenswrapper[4789]: I1008 14:03:11.733398 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:11 crc kubenswrapper[4789]: E1008 14:03:11.733485 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:03:12 crc kubenswrapper[4789]: I1008 14:03:12.729609 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:12 crc kubenswrapper[4789]: E1008 14:03:12.729821 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:03:13 crc kubenswrapper[4789]: I1008 14:03:13.768731 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:13 crc kubenswrapper[4789]: E1008 14:03:13.768892 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:03:13 crc kubenswrapper[4789]: I1008 14:03:13.768730 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:13 crc kubenswrapper[4789]: I1008 14:03:13.769117 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:13 crc kubenswrapper[4789]: E1008 14:03:13.769487 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:03:13 crc kubenswrapper[4789]: E1008 14:03:13.769567 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:03:14 crc kubenswrapper[4789]: I1008 14:03:14.729898 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:14 crc kubenswrapper[4789]: E1008 14:03:14.730170 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 14:03:15 crc kubenswrapper[4789]: I1008 14:03:15.729802 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:15 crc kubenswrapper[4789]: I1008 14:03:15.729806 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:15 crc kubenswrapper[4789]: E1008 14:03:15.730180 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b269s" podUID="765705a4-a303-4281-9677-5d0769f6d157" Oct 08 14:03:15 crc kubenswrapper[4789]: E1008 14:03:15.730005 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 14:03:15 crc kubenswrapper[4789]: I1008 14:03:15.729806 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:15 crc kubenswrapper[4789]: E1008 14:03:15.730312 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 14:03:16 crc kubenswrapper[4789]: I1008 14:03:16.730121 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:16 crc kubenswrapper[4789]: I1008 14:03:16.732477 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 08 14:03:16 crc kubenswrapper[4789]: I1008 14:03:16.733060 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 08 14:03:17 crc kubenswrapper[4789]: I1008 14:03:17.729920 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:17 crc kubenswrapper[4789]: I1008 14:03:17.730030 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:17 crc kubenswrapper[4789]: I1008 14:03:17.729920 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:17 crc kubenswrapper[4789]: I1008 14:03:17.731967 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 08 14:03:17 crc kubenswrapper[4789]: I1008 14:03:17.732075 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 08 14:03:17 crc kubenswrapper[4789]: I1008 14:03:17.732824 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 08 14:03:17 crc kubenswrapper[4789]: I1008 14:03:17.734799 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.768860 4789 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.811690 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rhnqk"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.812602 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.818005 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.818230 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.818289 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.818506 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.818762 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.818598 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.831453 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.832254 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.832485 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-config\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.832662 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.832800 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.832943 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9100571-6717-43fe-a0d8-0da2bd0ce31c-serving-cert\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.833197 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-client-ca\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.833392 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbzwl\" (UniqueName: \"kubernetes.io/projected/a9100571-6717-43fe-a0d8-0da2bd0ce31c-kube-api-access-xbzwl\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.832709 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.834439 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.834619 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.835342 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.838039 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kch4x"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.838719 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.839733 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.840694 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.841789 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.842479 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.842855 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.842869 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.843253 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.843521 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.843693 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.843742 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.843944 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.844070 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.848184 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nbtdd"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.858887 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.861924 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mtgx9"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.870020 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.906340 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.906931 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.907458 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.907689 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mvzv5"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.907774 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908105 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908161 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mvzv5" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908194 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908302 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908511 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908657 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908806 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908857 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908895 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.908970 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.909076 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.909175 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.909270 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.909365 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.909446 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.909509 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lkm5j"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.909540 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.910241 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-dwjgm"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.910640 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.910752 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.910782 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.910869 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.910937 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.911198 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.911317 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.911373 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.911527 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.911683 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.911861 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.912048 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.911224 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.912477 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.910897 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.915142 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.911272 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.911272 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.915367 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.915430 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.915495 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.915582 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.915587 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.915775 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.912089 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.918505 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.918911 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rmwkt"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.920114 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.921543 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.922128 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.930330 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.927701 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.930882 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wmqxc"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.928049 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.922210 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.931220 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.928245 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.928326 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.931808 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.931891 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.928327 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.928399 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.928492 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.928614 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.928692 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.929209 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.929319 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.929388 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.929516 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.929526 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.929542 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.934370 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmnn6\" (UniqueName: \"kubernetes.io/projected/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-kube-api-access-xmnn6\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.934655 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.934884 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e7dc58a5-148e-45bb-9895-a43fe46a0064-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935210 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935174 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935224 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-service-ca\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935469 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1764bb56-7b6e-4535-8af1-7d592163c9d1-config\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935503 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7dc58a5-148e-45bb-9895-a43fe46a0064-serving-cert\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935528 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5199b69-0cb4-4958-abaf-41089b4f0559-config\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935551 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk4kw\" (UniqueName: \"kubernetes.io/projected/bade0675-1189-4035-9673-2d20edbb729d-kube-api-access-hk4kw\") pod \"dns-operator-744455d44c-rmwkt\" (UID: \"bade0675-1189-4035-9673-2d20edbb729d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935601 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1764bb56-7b6e-4535-8af1-7d592163c9d1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935651 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e7dc58a5-148e-45bb-9895-a43fe46a0064-encryption-config\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935707 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935742 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-config\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935764 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-trusted-ca-bundle\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.935782 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.971426 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.971744 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7dc58a5-148e-45bb-9895-a43fe46a0064-audit-policies\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.971877 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.971953 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7dc58a5-148e-45bb-9895-a43fe46a0064-audit-dir\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.972071 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p8t9\" (UniqueName: \"kubernetes.io/projected/7191f22d-ffc0-47ff-b226-77f9213db94e-kube-api-access-4p8t9\") pod \"openshift-config-operator-7777fb866f-zmdbz\" (UID: \"7191f22d-ffc0-47ff-b226-77f9213db94e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.937430 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hf8hk"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.937537 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.938553 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.938760 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.940310 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.940616 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.940627 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.940741 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.940928 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.941547 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.978572 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-config\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.989469 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.989671 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.989742 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.989969 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.990175 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.990213 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.990242 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.990369 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.990447 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.990711 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.990783 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.990833 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.990941 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991047 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991106 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991185 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991357 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991367 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991560 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-p2mjq"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991648 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991422 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991705 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.991792 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993103 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-client-ca\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.978639 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-client-ca\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993172 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e7dc58a5-148e-45bb-9895-a43fe46a0064-etcd-client\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993189 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-audit\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993206 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60295e4c-18c7-4b3d-acbf-54964ed76081-config\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993220 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxld7\" (UniqueName: \"kubernetes.io/projected/6970b740-c77f-4817-b729-46839dc0698a-kube-api-access-vxld7\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993242 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993266 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58bece2-01c5-44eb-97ec-0de6757dd278-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lz74k\" (UID: \"d58bece2-01c5-44eb-97ec-0de6757dd278\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993281 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg7l8\" (UniqueName: \"kubernetes.io/projected/119eb4b6-56d0-4cae-803a-56020cda0096-kube-api-access-gg7l8\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993311 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhxkb\" (UniqueName: \"kubernetes.io/projected/d58bece2-01c5-44eb-97ec-0de6757dd278-kube-api-access-hhxkb\") pod \"openshift-apiserver-operator-796bbdcf4f-lz74k\" (UID: \"d58bece2-01c5-44eb-97ec-0de6757dd278\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993326 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drd2q\" (UniqueName: \"kubernetes.io/projected/e7dc58a5-148e-45bb-9895-a43fe46a0064-kube-api-access-drd2q\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993345 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1764bb56-7b6e-4535-8af1-7d592163c9d1-serving-cert\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993363 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p78jx\" (UniqueName: \"kubernetes.io/projected/53dc6d10-02c4-48c5-8461-85ed095bcc11-kube-api-access-p78jx\") pod \"cluster-samples-operator-665b6dd947-7pwrl\" (UID: \"53dc6d10-02c4-48c5-8461-85ed095bcc11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993380 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993396 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-config\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993418 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993435 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-serving-cert\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993452 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993467 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/119eb4b6-56d0-4cae-803a-56020cda0096-audit-dir\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993483 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993500 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6970b740-c77f-4817-b729-46839dc0698a-node-pullsecrets\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993520 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b5199b69-0cb4-4958-abaf-41089b4f0559-trusted-ca\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993537 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9100571-6717-43fe-a0d8-0da2bd0ce31c-serving-cert\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993551 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6970b740-c77f-4817-b729-46839dc0698a-serving-cert\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993566 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwgzr\" (UniqueName: \"kubernetes.io/projected/b5199b69-0cb4-4958-abaf-41089b4f0559-kube-api-access-qwgzr\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993587 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-image-import-ca\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993601 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-audit-policies\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993615 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-client-ca\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993629 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5199b69-0cb4-4958-abaf-41089b4f0559-serving-cert\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993655 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d58bece2-01c5-44eb-97ec-0de6757dd278-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lz74k\" (UID: \"d58bece2-01c5-44eb-97ec-0de6757dd278\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993668 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/60295e4c-18c7-4b3d-acbf-54964ed76081-auth-proxy-config\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993684 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993701 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-oauth-config\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993716 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6970b740-c77f-4817-b729-46839dc0698a-audit-dir\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993730 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-oauth-serving-cert\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993753 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-etcd-serving-ca\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993768 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh58c\" (UniqueName: \"kubernetes.io/projected/60295e4c-18c7-4b3d-acbf-54964ed76081-kube-api-access-rh58c\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993783 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993799 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993815 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993833 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6970b740-c77f-4817-b729-46839dc0698a-etcd-client\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993857 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/60295e4c-18c7-4b3d-acbf-54964ed76081-machine-approver-tls\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993872 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hssvh\" (UniqueName: \"kubernetes.io/projected/6fb689b3-3b0d-476a-ba26-f6836061f366-kube-api-access-hssvh\") pod \"downloads-7954f5f757-mvzv5\" (UID: \"6fb689b3-3b0d-476a-ba26-f6836061f366\") " pod="openshift-console/downloads-7954f5f757-mvzv5" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993886 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-serving-cert\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993902 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993925 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/53dc6d10-02c4-48c5-8461-85ed095bcc11-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7pwrl\" (UID: \"53dc6d10-02c4-48c5-8461-85ed095bcc11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993942 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7191f22d-ffc0-47ff-b226-77f9213db94e-serving-cert\") pod \"openshift-config-operator-7777fb866f-zmdbz\" (UID: \"7191f22d-ffc0-47ff-b226-77f9213db94e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993959 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwjzz\" (UniqueName: \"kubernetes.io/projected/6fd06952-1b3e-44e6-af19-dfa31965b9ea-kube-api-access-zwjzz\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.993974 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7191f22d-ffc0-47ff-b226-77f9213db94e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zmdbz\" (UID: \"7191f22d-ffc0-47ff-b226-77f9213db94e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994005 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1764bb56-7b6e-4535-8af1-7d592163c9d1-service-ca-bundle\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994020 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994036 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6dgm\" (UniqueName: \"kubernetes.io/projected/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-kube-api-access-l6dgm\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994055 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbzwl\" (UniqueName: \"kubernetes.io/projected/a9100571-6717-43fe-a0d8-0da2bd0ce31c-kube-api-access-xbzwl\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994072 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7dc58a5-148e-45bb-9895-a43fe46a0064-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994085 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bade0675-1189-4035-9673-2d20edbb729d-metrics-tls\") pod \"dns-operator-744455d44c-rmwkt\" (UID: \"bade0675-1189-4035-9673-2d20edbb729d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994101 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-config\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994116 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz749\" (UniqueName: \"kubernetes.io/projected/1764bb56-7b6e-4535-8af1-7d592163c9d1-kube-api-access-nz749\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994134 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6970b740-c77f-4817-b729-46839dc0698a-encryption-config\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994149 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994310 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mc249"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994621 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.994891 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.995008 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vzq6g"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.996069 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.996181 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.996248 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.996390 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.996519 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-config\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.996592 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.997054 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.997101 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.997244 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.998026 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xs8k5"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.998241 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.998494 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.998595 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.998760 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls"] Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.998816 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:19 crc kubenswrapper[4789]: I1008 14:03:19.999184 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.001100 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.001954 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-td8vc"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.002807 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.002945 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.003129 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.003834 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.004246 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.004569 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.005605 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.007175 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.007625 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.008206 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.010794 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9100571-6717-43fe-a0d8-0da2bd0ce31c-serving-cert\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.013195 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.015218 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.015373 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.015472 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.015561 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.016774 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.016849 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.020187 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.025095 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.031783 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.034863 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.041060 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rmwkt"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.044616 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.053213 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.053247 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-jgbjw"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.054202 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.055127 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-dwjgm"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.055221 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.056551 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.057464 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.059129 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mtgx9"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.060837 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mvzv5"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.061864 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.064000 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hf8hk"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.066090 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.067363 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xs8k5"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.068661 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mc249"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.069904 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vzq6g"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.071477 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wmqxc"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.073321 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lkm5j"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.075034 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.075069 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.076197 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.077572 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.079335 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rhnqk"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.080810 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nbtdd"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.081896 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.083000 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.084102 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.086245 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kch4x"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.089823 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.091462 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.093245 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-96vg6"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.094399 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.094511 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.094759 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg4pz\" (UniqueName: \"kubernetes.io/projected/bc2cc553-6ac7-44a6-948c-c1363836923d-kube-api-access-bg4pz\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.094897 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09101a18-6c60-4da9-b517-afb6ae89cb4c-service-ca-bundle\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095006 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6970b740-c77f-4817-b729-46839dc0698a-serving-cert\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095100 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/449f550c-38c6-4c62-abc7-2566c2cf1ccc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l7dth\" (UID: \"449f550c-38c6-4c62-abc7-2566c2cf1ccc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095207 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095207 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-image-import-ca\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-audit-policies\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095298 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-client-ca\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095323 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwgzr\" (UniqueName: \"kubernetes.io/projected/b5199b69-0cb4-4958-abaf-41089b4f0559-kube-api-access-qwgzr\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095365 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095391 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5199b69-0cb4-4958-abaf-41089b4f0559-serving-cert\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095420 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-serving-cert\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095442 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vzq6g\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095472 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-etcd-serving-ca\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095496 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-oauth-serving-cert\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095520 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh58c\" (UniqueName: \"kubernetes.io/projected/60295e4c-18c7-4b3d-acbf-54964ed76081-kube-api-access-rh58c\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095543 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095567 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095589 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/60295e4c-18c7-4b3d-acbf-54964ed76081-machine-approver-tls\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095614 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc2cc553-6ac7-44a6-948c-c1363836923d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095636 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/09101a18-6c60-4da9-b517-afb6ae89cb4c-default-certificate\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095659 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-proxy-tls\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095683 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-serving-cert\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095705 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095728 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095752 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-config\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095777 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7191f22d-ffc0-47ff-b226-77f9213db94e-serving-cert\") pod \"openshift-config-operator-7777fb866f-zmdbz\" (UID: \"7191f22d-ffc0-47ff-b226-77f9213db94e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095808 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1764bb56-7b6e-4535-8af1-7d592163c9d1-service-ca-bundle\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095831 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7dc58a5-148e-45bb-9895-a43fe46a0064-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095861 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-webhook-cert\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095885 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8317867-1744-48e0-8940-35c3a0c2ab1b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-g9tdd\" (UID: \"f8317867-1744-48e0-8940-35c3a0c2ab1b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095915 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e7dc58a5-148e-45bb-9895-a43fe46a0064-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095940 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09101a18-6c60-4da9-b517-afb6ae89cb4c-metrics-certs\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.095965 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cbd6f891-627b-4f0b-ac4d-40e30d01fb89-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c8rg8\" (UID: \"cbd6f891-627b-4f0b-ac4d-40e30d01fb89\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096007 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8317867-1744-48e0-8940-35c3a0c2ab1b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-g9tdd\" (UID: \"f8317867-1744-48e0-8940-35c3a0c2ab1b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096031 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7dc58a5-148e-45bb-9895-a43fe46a0064-serving-cert\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096054 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5199b69-0cb4-4958-abaf-41089b4f0559-config\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096077 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7rt7\" (UniqueName: \"kubernetes.io/projected/73df2e86-6ba0-47d8-bfac-ef6d675808e6-kube-api-access-p7rt7\") pod \"package-server-manager-789f6589d5-hzztl\" (UID: \"73df2e86-6ba0-47d8-bfac-ef6d675808e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096100 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1764bb56-7b6e-4535-8af1-7d592163c9d1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096121 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-images\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096144 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-config\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096167 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-trusted-ca-bundle\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096191 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1c546fd5-9ea3-402e-b837-dca47372c52c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rbvwr\" (UID: \"1c546fd5-9ea3-402e-b837-dca47372c52c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096194 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-client-ca\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096215 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m7tl\" (UniqueName: \"kubernetes.io/projected/f8317867-1744-48e0-8940-35c3a0c2ab1b-kube-api-access-6m7tl\") pod \"openshift-controller-manager-operator-756b6f6bc6-g9tdd\" (UID: \"f8317867-1744-48e0-8940-35c3a0c2ab1b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096362 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7dc58a5-148e-45bb-9895-a43fe46a0064-audit-policies\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096378 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-audit-policies\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096387 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096412 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-apiservice-cert\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096436 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa675ad1-38c1-49b7-9572-f165efdbd2e9-trusted-ca\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096459 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6kxj\" (UniqueName: \"kubernetes.io/projected/aa675ad1-38c1-49b7-9572-f165efdbd2e9-kube-api-access-m6kxj\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096489 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-config\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096511 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60295e4c-18c7-4b3d-acbf-54964ed76081-config\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096537 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096560 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2389a066-e146-4ea8-bb92-d596311b564a-secret-volume\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096583 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17ebe9cd-3f39-42c3-a888-7270e67caa06-config\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096608 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58bece2-01c5-44eb-97ec-0de6757dd278-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lz74k\" (UID: \"d58bece2-01c5-44eb-97ec-0de6757dd278\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096631 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg7l8\" (UniqueName: \"kubernetes.io/projected/119eb4b6-56d0-4cae-803a-56020cda0096-kube-api-access-gg7l8\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096654 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnps4\" (UniqueName: \"kubernetes.io/projected/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-kube-api-access-lnps4\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096678 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ae8586c-f105-42ff-b845-b4eb6a878694-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dqrvj\" (UID: \"6ae8586c-f105-42ff-b845-b4eb6a878694\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096698 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/17ebe9cd-3f39-42c3-a888-7270e67caa06-etcd-ca\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096738 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhxkb\" (UniqueName: \"kubernetes.io/projected/d58bece2-01c5-44eb-97ec-0de6757dd278-kube-api-access-hhxkb\") pod \"openshift-apiserver-operator-796bbdcf4f-lz74k\" (UID: \"d58bece2-01c5-44eb-97ec-0de6757dd278\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096764 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/73df2e86-6ba0-47d8-bfac-ef6d675808e6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hzztl\" (UID: \"73df2e86-6ba0-47d8-bfac-ef6d675808e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096793 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-config\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096818 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.096976 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-gr58t"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.097736 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.098609 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.098946 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e7dc58a5-148e-45bb-9895-a43fe46a0064-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.099055 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6970b740-c77f-4817-b729-46839dc0698a-serving-cert\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.099187 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.099199 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.099270 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-etcd-serving-ca\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.099335 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/60295e4c-18c7-4b3d-acbf-54964ed76081-machine-approver-tls\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.099705 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1764bb56-7b6e-4535-8af1-7d592163c9d1-service-ca-bundle\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.099739 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.099904 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-config\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.100090 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1764bb56-7b6e-4535-8af1-7d592163c9d1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.100407 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7dc58a5-148e-45bb-9895-a43fe46a0064-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.100548 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/119eb4b6-56d0-4cae-803a-56020cda0096-audit-dir\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.100759 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-config\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101040 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58bece2-01c5-44eb-97ec-0de6757dd278-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lz74k\" (UID: \"d58bece2-01c5-44eb-97ec-0de6757dd278\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.098096 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/119eb4b6-56d0-4cae-803a-56020cda0096-audit-dir\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101166 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5199b69-0cb4-4958-abaf-41089b4f0559-serving-cert\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101285 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101375 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b5199b69-0cb4-4958-abaf-41089b4f0559-trusted-ca\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101427 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/09101a18-6c60-4da9-b517-afb6ae89cb4c-stats-auth\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101447 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv4wp\" (UniqueName: \"kubernetes.io/projected/1c546fd5-9ea3-402e-b837-dca47372c52c-kube-api-access-wv4wp\") pod \"machine-config-controller-84d6567774-rbvwr\" (UID: \"1c546fd5-9ea3-402e-b837-dca47372c52c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101475 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bc2cc553-6ac7-44a6-948c-c1363836923d-images\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101493 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2389a066-e146-4ea8-bb92-d596311b564a-config-volume\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101513 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6jt8\" (UniqueName: \"kubernetes.io/projected/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-kube-api-access-r6jt8\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101546 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/17ebe9cd-3f39-42c3-a888-7270e67caa06-etcd-service-ca\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101567 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d58bece2-01c5-44eb-97ec-0de6757dd278-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lz74k\" (UID: \"d58bece2-01c5-44eb-97ec-0de6757dd278\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101595 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/60295e4c-18c7-4b3d-acbf-54964ed76081-auth-proxy-config\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101614 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-oauth-config\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101649 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6970b740-c77f-4817-b729-46839dc0698a-audit-dir\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101665 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101665 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-serving-cert\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101683 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101703 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6970b740-c77f-4817-b729-46839dc0698a-etcd-client\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101701 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e7dc58a5-148e-45bb-9895-a43fe46a0064-audit-policies\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101739 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hssvh\" (UniqueName: \"kubernetes.io/projected/6fb689b3-3b0d-476a-ba26-f6836061f366-kube-api-access-hssvh\") pod \"downloads-7954f5f757-mvzv5\" (UID: \"6fb689b3-3b0d-476a-ba26-f6836061f366\") " pod="openshift-console/downloads-7954f5f757-mvzv5" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101805 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/449f550c-38c6-4c62-abc7-2566c2cf1ccc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l7dth\" (UID: \"449f550c-38c6-4c62-abc7-2566c2cf1ccc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101827 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7jpl\" (UniqueName: \"kubernetes.io/projected/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-kube-api-access-x7jpl\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101846 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa675ad1-38c1-49b7-9572-f165efdbd2e9-metrics-tls\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101874 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/53dc6d10-02c4-48c5-8461-85ed095bcc11-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7pwrl\" (UID: \"53dc6d10-02c4-48c5-8461-85ed095bcc11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101916 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwjzz\" (UniqueName: \"kubernetes.io/projected/6fd06952-1b3e-44e6-af19-dfa31965b9ea-kube-api-access-zwjzz\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101934 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7191f22d-ffc0-47ff-b226-77f9213db94e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zmdbz\" (UID: \"7191f22d-ffc0-47ff-b226-77f9213db94e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101950 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.101965 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6dgm\" (UniqueName: \"kubernetes.io/projected/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-kube-api-access-l6dgm\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102012 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h9lt\" (UniqueName: \"kubernetes.io/projected/5b3f76e2-8cce-47f1-87e5-43be0d07238e-kube-api-access-2h9lt\") pod \"marketplace-operator-79b997595-vzq6g\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102098 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-cabundle\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102122 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bade0675-1189-4035-9673-2d20edbb729d-metrics-tls\") pod \"dns-operator-744455d44c-rmwkt\" (UID: \"bade0675-1189-4035-9673-2d20edbb729d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102146 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1c546fd5-9ea3-402e-b837-dca47372c52c-proxy-tls\") pod \"machine-config-controller-84d6567774-rbvwr\" (UID: \"1c546fd5-9ea3-402e-b837-dca47372c52c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102162 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlj85\" (UniqueName: \"kubernetes.io/projected/09101a18-6c60-4da9-b517-afb6ae89cb4c-kube-api-access-nlj85\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102177 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17ebe9cd-3f39-42c3-a888-7270e67caa06-serving-cert\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102198 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz749\" (UniqueName: \"kubernetes.io/projected/1764bb56-7b6e-4535-8af1-7d592163c9d1-kube-api-access-nz749\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102213 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6970b740-c77f-4817-b729-46839dc0698a-encryption-config\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102228 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102244 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc2cc553-6ac7-44a6-948c-c1363836923d-config\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102260 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-key\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102812 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.102839 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60295e4c-18c7-4b3d-acbf-54964ed76081-config\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.103034 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-image-import-ca\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.103193 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-config\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.103235 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7dc58a5-148e-45bb-9895-a43fe46a0064-serving-cert\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.103430 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-oauth-serving-cert\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.103508 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.103611 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-trusted-ca-bundle\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.103872 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmnn6\" (UniqueName: \"kubernetes.io/projected/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-kube-api-access-xmnn6\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.103930 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b5199b69-0cb4-4958-abaf-41089b4f0559-trusted-ca\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.103975 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6970b740-c77f-4817-b729-46839dc0698a-audit-dir\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104007 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104024 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104189 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-service-ca\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104200 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104231 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vzq6g\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104287 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1764bb56-7b6e-4535-8af1-7d592163c9d1-config\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104325 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk4kw\" (UniqueName: \"kubernetes.io/projected/bade0675-1189-4035-9673-2d20edbb729d-kube-api-access-hk4kw\") pod \"dns-operator-744455d44c-rmwkt\" (UID: \"bade0675-1189-4035-9673-2d20edbb729d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104460 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-td8vc"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104482 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/60295e4c-18c7-4b3d-acbf-54964ed76081-auth-proxy-config\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104499 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e7dc58a5-148e-45bb-9895-a43fe46a0064-encryption-config\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104530 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/449f550c-38c6-4c62-abc7-2566c2cf1ccc-config\") pod \"kube-controller-manager-operator-78b949d7b-l7dth\" (UID: \"449f550c-38c6-4c62-abc7-2566c2cf1ccc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104566 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104591 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104618 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae8586c-f105-42ff-b845-b4eb6a878694-config\") pod \"kube-apiserver-operator-766d6c64bb-dqrvj\" (UID: \"6ae8586c-f105-42ff-b845-b4eb6a878694\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104641 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b9ht\" (UniqueName: \"kubernetes.io/projected/cbd6f891-627b-4f0b-ac4d-40e30d01fb89-kube-api-access-9b9ht\") pod \"control-plane-machine-set-operator-78cbb6b69f-c8rg8\" (UID: \"cbd6f891-627b-4f0b-ac4d-40e30d01fb89\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104679 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r67d6\" (UniqueName: \"kubernetes.io/projected/2389a066-e146-4ea8-bb92-d596311b564a-kube-api-access-r67d6\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104704 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104729 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wvkn\" (UniqueName: \"kubernetes.io/projected/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-kube-api-access-2wvkn\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104751 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa675ad1-38c1-49b7-9572-f165efdbd2e9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104776 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p8t9\" (UniqueName: \"kubernetes.io/projected/7191f22d-ffc0-47ff-b226-77f9213db94e-kube-api-access-4p8t9\") pod \"openshift-config-operator-7777fb866f-zmdbz\" (UID: \"7191f22d-ffc0-47ff-b226-77f9213db94e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104783 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-service-ca\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104799 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7dc58a5-148e-45bb-9895-a43fe46a0064-audit-dir\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104827 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/17ebe9cd-3f39-42c3-a888-7270e67caa06-etcd-client\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104855 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e7dc58a5-148e-45bb-9895-a43fe46a0064-audit-dir\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.104901 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1764bb56-7b6e-4535-8af1-7d592163c9d1-config\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105166 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e7dc58a5-148e-45bb-9895-a43fe46a0064-etcd-client\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105172 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d58bece2-01c5-44eb-97ec-0de6757dd278-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lz74k\" (UID: \"d58bece2-01c5-44eb-97ec-0de6757dd278\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105214 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-audit\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105250 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6psjr\" (UniqueName: \"kubernetes.io/projected/17ebe9cd-3f39-42c3-a888-7270e67caa06-kube-api-access-6psjr\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105363 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxld7\" (UniqueName: \"kubernetes.io/projected/6970b740-c77f-4817-b729-46839dc0698a-kube-api-access-vxld7\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105518 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crwnl\" (UniqueName: \"kubernetes.io/projected/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-kube-api-access-crwnl\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105563 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ae8586c-f105-42ff-b845-b4eb6a878694-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dqrvj\" (UID: \"6ae8586c-f105-42ff-b845-b4eb6a878694\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105597 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drd2q\" (UniqueName: \"kubernetes.io/projected/e7dc58a5-148e-45bb-9895-a43fe46a0064-kube-api-access-drd2q\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105736 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1764bb56-7b6e-4535-8af1-7d592163c9d1-serving-cert\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.105747 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-audit\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106033 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6970b740-c77f-4817-b729-46839dc0698a-etcd-client\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106182 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5199b69-0cb4-4958-abaf-41089b4f0559-config\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106208 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106314 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7191f22d-ffc0-47ff-b226-77f9213db94e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zmdbz\" (UID: \"7191f22d-ffc0-47ff-b226-77f9213db94e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106348 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-srv-cert\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106486 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-tmpfs\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106581 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p78jx\" (UniqueName: \"kubernetes.io/projected/53dc6d10-02c4-48c5-8461-85ed095bcc11-kube-api-access-p78jx\") pod \"cluster-samples-operator-665b6dd947-7pwrl\" (UID: \"53dc6d10-02c4-48c5-8461-85ed095bcc11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106662 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106689 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-oauth-config\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106514 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106745 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-serving-cert\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106833 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6970b740-c77f-4817-b729-46839dc0698a-node-pullsecrets\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.106975 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6970b740-c77f-4817-b729-46839dc0698a-node-pullsecrets\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.107139 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-96vg6"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.107383 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6970b740-c77f-4817-b729-46839dc0698a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.107753 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bade0675-1189-4035-9673-2d20edbb729d-metrics-tls\") pod \"dns-operator-744455d44c-rmwkt\" (UID: \"bade0675-1189-4035-9673-2d20edbb729d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.108331 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1764bb56-7b6e-4535-8af1-7d592163c9d1-serving-cert\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.108692 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.109101 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.109230 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e7dc58a5-148e-45bb-9895-a43fe46a0064-etcd-client\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.109367 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6970b740-c77f-4817-b729-46839dc0698a-encryption-config\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.110536 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.110832 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-serving-cert\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.111739 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.111757 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e7dc58a5-148e-45bb-9895-a43fe46a0064-encryption-config\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.111833 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gr58t"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.111878 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.112102 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.112242 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/53dc6d10-02c4-48c5-8461-85ed095bcc11-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7pwrl\" (UID: \"53dc6d10-02c4-48c5-8461-85ed095bcc11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.114912 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.115675 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.116348 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.118450 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-dncbh"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.119829 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dncbh" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.120433 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dncbh"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.126433 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7191f22d-ffc0-47ff-b226-77f9213db94e-serving-cert\") pod \"openshift-config-operator-7777fb866f-zmdbz\" (UID: \"7191f22d-ffc0-47ff-b226-77f9213db94e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.135560 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.154713 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.175015 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.195331 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.207916 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09101a18-6c60-4da9-b517-afb6ae89cb4c-service-ca-bundle\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.207976 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg4pz\" (UniqueName: \"kubernetes.io/projected/bc2cc553-6ac7-44a6-948c-c1363836923d-kube-api-access-bg4pz\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208011 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/449f550c-38c6-4c62-abc7-2566c2cf1ccc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l7dth\" (UID: \"449f550c-38c6-4c62-abc7-2566c2cf1ccc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208035 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-serving-cert\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208057 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vzq6g\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208083 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc2cc553-6ac7-44a6-948c-c1363836923d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208098 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/09101a18-6c60-4da9-b517-afb6ae89cb4c-default-certificate\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208116 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-proxy-tls\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208134 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208149 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-config\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208165 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8317867-1744-48e0-8940-35c3a0c2ab1b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-g9tdd\" (UID: \"f8317867-1744-48e0-8940-35c3a0c2ab1b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208182 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-webhook-cert\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208199 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09101a18-6c60-4da9-b517-afb6ae89cb4c-metrics-certs\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208216 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cbd6f891-627b-4f0b-ac4d-40e30d01fb89-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c8rg8\" (UID: \"cbd6f891-627b-4f0b-ac4d-40e30d01fb89\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208233 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8317867-1744-48e0-8940-35c3a0c2ab1b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-g9tdd\" (UID: \"f8317867-1744-48e0-8940-35c3a0c2ab1b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208252 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7rt7\" (UniqueName: \"kubernetes.io/projected/73df2e86-6ba0-47d8-bfac-ef6d675808e6-kube-api-access-p7rt7\") pod \"package-server-manager-789f6589d5-hzztl\" (UID: \"73df2e86-6ba0-47d8-bfac-ef6d675808e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208269 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-images\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208289 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m7tl\" (UniqueName: \"kubernetes.io/projected/f8317867-1744-48e0-8940-35c3a0c2ab1b-kube-api-access-6m7tl\") pod \"openshift-controller-manager-operator-756b6f6bc6-g9tdd\" (UID: \"f8317867-1744-48e0-8940-35c3a0c2ab1b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208308 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1c546fd5-9ea3-402e-b837-dca47372c52c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rbvwr\" (UID: \"1c546fd5-9ea3-402e-b837-dca47372c52c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208323 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-apiservice-cert\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208338 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa675ad1-38c1-49b7-9572-f165efdbd2e9-trusted-ca\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208355 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6kxj\" (UniqueName: \"kubernetes.io/projected/aa675ad1-38c1-49b7-9572-f165efdbd2e9-kube-api-access-m6kxj\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208374 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2389a066-e146-4ea8-bb92-d596311b564a-secret-volume\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208391 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17ebe9cd-3f39-42c3-a888-7270e67caa06-config\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208412 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnps4\" (UniqueName: \"kubernetes.io/projected/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-kube-api-access-lnps4\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208428 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ae8586c-f105-42ff-b845-b4eb6a878694-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dqrvj\" (UID: \"6ae8586c-f105-42ff-b845-b4eb6a878694\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208441 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/17ebe9cd-3f39-42c3-a888-7270e67caa06-etcd-ca\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208470 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/73df2e86-6ba0-47d8-bfac-ef6d675808e6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hzztl\" (UID: \"73df2e86-6ba0-47d8-bfac-ef6d675808e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208487 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/09101a18-6c60-4da9-b517-afb6ae89cb4c-stats-auth\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208510 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv4wp\" (UniqueName: \"kubernetes.io/projected/1c546fd5-9ea3-402e-b837-dca47372c52c-kube-api-access-wv4wp\") pod \"machine-config-controller-84d6567774-rbvwr\" (UID: \"1c546fd5-9ea3-402e-b837-dca47372c52c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208528 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bc2cc553-6ac7-44a6-948c-c1363836923d-images\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208545 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2389a066-e146-4ea8-bb92-d596311b564a-config-volume\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208560 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/17ebe9cd-3f39-42c3-a888-7270e67caa06-etcd-service-ca\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208577 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6jt8\" (UniqueName: \"kubernetes.io/projected/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-kube-api-access-r6jt8\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208600 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208615 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/449f550c-38c6-4c62-abc7-2566c2cf1ccc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l7dth\" (UID: \"449f550c-38c6-4c62-abc7-2566c2cf1ccc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208631 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7jpl\" (UniqueName: \"kubernetes.io/projected/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-kube-api-access-x7jpl\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208649 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa675ad1-38c1-49b7-9572-f165efdbd2e9-metrics-tls\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208689 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h9lt\" (UniqueName: \"kubernetes.io/projected/5b3f76e2-8cce-47f1-87e5-43be0d07238e-kube-api-access-2h9lt\") pod \"marketplace-operator-79b997595-vzq6g\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208706 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-cabundle\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208722 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1c546fd5-9ea3-402e-b837-dca47372c52c-proxy-tls\") pod \"machine-config-controller-84d6567774-rbvwr\" (UID: \"1c546fd5-9ea3-402e-b837-dca47372c52c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208742 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlj85\" (UniqueName: \"kubernetes.io/projected/09101a18-6c60-4da9-b517-afb6ae89cb4c-kube-api-access-nlj85\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208759 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17ebe9cd-3f39-42c3-a888-7270e67caa06-serving-cert\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208781 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc2cc553-6ac7-44a6-948c-c1363836923d-config\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208795 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-key\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208816 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vzq6g\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208837 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/449f550c-38c6-4c62-abc7-2566c2cf1ccc-config\") pod \"kube-controller-manager-operator-78b949d7b-l7dth\" (UID: \"449f550c-38c6-4c62-abc7-2566c2cf1ccc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208853 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae8586c-f105-42ff-b845-b4eb6a878694-config\") pod \"kube-apiserver-operator-766d6c64bb-dqrvj\" (UID: \"6ae8586c-f105-42ff-b845-b4eb6a878694\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208870 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b9ht\" (UniqueName: \"kubernetes.io/projected/cbd6f891-627b-4f0b-ac4d-40e30d01fb89-kube-api-access-9b9ht\") pod \"control-plane-machine-set-operator-78cbb6b69f-c8rg8\" (UID: \"cbd6f891-627b-4f0b-ac4d-40e30d01fb89\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208895 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r67d6\" (UniqueName: \"kubernetes.io/projected/2389a066-e146-4ea8-bb92-d596311b564a-kube-api-access-r67d6\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208912 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wvkn\" (UniqueName: \"kubernetes.io/projected/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-kube-api-access-2wvkn\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208927 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa675ad1-38c1-49b7-9572-f165efdbd2e9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208948 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/17ebe9cd-3f39-42c3-a888-7270e67caa06-etcd-client\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.208966 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6psjr\" (UniqueName: \"kubernetes.io/projected/17ebe9cd-3f39-42c3-a888-7270e67caa06-kube-api-access-6psjr\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209016 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crwnl\" (UniqueName: \"kubernetes.io/projected/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-kube-api-access-crwnl\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209036 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ae8586c-f105-42ff-b845-b4eb6a878694-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dqrvj\" (UID: \"6ae8586c-f105-42ff-b845-b4eb6a878694\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209057 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-srv-cert\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209072 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-tmpfs\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209188 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1c546fd5-9ea3-402e-b837-dca47372c52c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rbvwr\" (UID: \"1c546fd5-9ea3-402e-b837-dca47372c52c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209365 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8317867-1744-48e0-8940-35c3a0c2ab1b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-g9tdd\" (UID: \"f8317867-1744-48e0-8940-35c3a0c2ab1b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209424 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209549 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-tmpfs\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209629 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bc2cc553-6ac7-44a6-948c-c1363836923d-images\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.209780 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc2cc553-6ac7-44a6-948c-c1363836923d-config\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.213618 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc2cc553-6ac7-44a6-948c-c1363836923d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.213974 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa675ad1-38c1-49b7-9572-f165efdbd2e9-metrics-tls\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.214888 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8317867-1744-48e0-8940-35c3a0c2ab1b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-g9tdd\" (UID: \"f8317867-1744-48e0-8940-35c3a0c2ab1b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.220226 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.230101 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa675ad1-38c1-49b7-9572-f165efdbd2e9-trusted-ca\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.234235 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.275041 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.279459 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/17ebe9cd-3f39-42c3-a888-7270e67caa06-etcd-ca\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.294385 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.300137 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17ebe9cd-3f39-42c3-a888-7270e67caa06-config\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.315149 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.334324 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.339263 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/17ebe9cd-3f39-42c3-a888-7270e67caa06-etcd-service-ca\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.354768 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.362592 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17ebe9cd-3f39-42c3-a888-7270e67caa06-serving-cert\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.375350 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.395492 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.405951 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/17ebe9cd-3f39-42c3-a888-7270e67caa06-etcd-client\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.416898 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.435346 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.455884 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.475664 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.483470 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/449f550c-38c6-4c62-abc7-2566c2cf1ccc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l7dth\" (UID: \"449f550c-38c6-4c62-abc7-2566c2cf1ccc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.494908 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.500702 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/449f550c-38c6-4c62-abc7-2566c2cf1ccc-config\") pod \"kube-controller-manager-operator-78b949d7b-l7dth\" (UID: \"449f550c-38c6-4c62-abc7-2566c2cf1ccc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.515727 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.534967 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.540890 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae8586c-f105-42ff-b845-b4eb6a878694-config\") pod \"kube-apiserver-operator-766d6c64bb-dqrvj\" (UID: \"6ae8586c-f105-42ff-b845-b4eb6a878694\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.556089 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.563074 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ae8586c-f105-42ff-b845-b4eb6a878694-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dqrvj\" (UID: \"6ae8586c-f105-42ff-b845-b4eb6a878694\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.575962 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.616460 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.622813 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbzwl\" (UniqueName: \"kubernetes.io/projected/a9100571-6717-43fe-a0d8-0da2bd0ce31c-kube-api-access-xbzwl\") pod \"controller-manager-879f6c89f-rhnqk\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.623501 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1c546fd5-9ea3-402e-b837-dca47372c52c-proxy-tls\") pod \"machine-config-controller-84d6567774-rbvwr\" (UID: \"1c546fd5-9ea3-402e-b837-dca47372c52c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.635836 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.654683 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.674836 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.695742 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.715580 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.735134 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.748699 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.755324 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.775568 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.779238 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vzq6g\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.782297 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/09101a18-6c60-4da9-b517-afb6ae89cb4c-default-certificate\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.795388 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.816807 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.829470 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/09101a18-6c60-4da9-b517-afb6ae89cb4c-stats-auth\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.835027 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.855719 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.885464 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.892177 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vzq6g\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.895541 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.899091 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09101a18-6c60-4da9-b517-afb6ae89cb4c-service-ca-bundle\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.915181 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.933536 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rhnqk"] Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.934679 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 08 14:03:20 crc kubenswrapper[4789]: W1008 14:03:20.941318 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9100571_6717_43fe_a0d8_0da2bd0ce31c.slice/crio-54be8b8f18e145300894a81e3624eb13ca8a917defdee2c30c3708705bd12c4f WatchSource:0}: Error finding container 54be8b8f18e145300894a81e3624eb13ca8a917defdee2c30c3708705bd12c4f: Status 404 returned error can't find the container with id 54be8b8f18e145300894a81e3624eb13ca8a917defdee2c30c3708705bd12c4f Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.943050 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09101a18-6c60-4da9-b517-afb6ae89cb4c-metrics-certs\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.954837 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.963433 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-images\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.975124 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 08 14:03:20 crc kubenswrapper[4789]: I1008 14:03:20.995269 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.002931 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-proxy-tls\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.014229 4789 request.go:700] Waited for 1.015780065s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-operator-dockercfg-2bh8d&limit=500&resourceVersion=0 Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.017430 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.036598 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.055304 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.075196 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.095425 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.115925 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.134887 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.155516 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.175158 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.194952 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.202723 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-apiservice-cert\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.204543 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-webhook-cert\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209158 4789 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209231 4789 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209240 4789 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209261 4789 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209285 4789 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209245 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cbd6f891-627b-4f0b-ac4d-40e30d01fb89-control-plane-machine-set-operator-tls podName:cbd6f891-627b-4f0b-ac4d-40e30d01fb89 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.709219301 +0000 UTC m=+141.615966883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/cbd6f891-627b-4f0b-ac4d-40e30d01fb89-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-c8rg8" (UID: "cbd6f891-627b-4f0b-ac4d-40e30d01fb89") : failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209250 4789 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209342 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73df2e86-6ba0-47d8-bfac-ef6d675808e6-package-server-manager-serving-cert podName:73df2e86-6ba0-47d8-bfac-ef6d675808e6 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.709324444 +0000 UTC m=+141.616071946 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/73df2e86-6ba0-47d8-bfac-ef6d675808e6-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-hzztl" (UID: "73df2e86-6ba0-47d8-bfac-ef6d675808e6") : failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209362 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-config podName:8e83dccc-9fa2-42d5-ae9a-7941ae00d321 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.709351344 +0000 UTC m=+141.616098846 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-config") pod "service-ca-operator-777779d784-kx5ls" (UID: "8e83dccc-9fa2-42d5-ae9a-7941ae00d321") : failed to sync configmap cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209378 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-serving-cert podName:8e83dccc-9fa2-42d5-ae9a-7941ae00d321 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.709372055 +0000 UTC m=+141.616119557 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-serving-cert") pod "service-ca-operator-777779d784-kx5ls" (UID: "8e83dccc-9fa2-42d5-ae9a-7941ae00d321") : failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209405 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2389a066-e146-4ea8-bb92-d596311b564a-secret-volume podName:2389a066-e146-4ea8-bb92-d596311b564a nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.709397606 +0000 UTC m=+141.616145108 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/2389a066-e146-4ea8-bb92-d596311b564a-secret-volume") pod "collect-profiles-29332200-8zphp" (UID: "2389a066-e146-4ea8-bb92-d596311b564a") : failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.209429 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-profile-collector-cert podName:4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.709419116 +0000 UTC m=+141.616166618 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-profile-collector-cert") pod "catalog-operator-68c6474976-bkblm" (UID: "4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9") : failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.210480 4789 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.210499 4789 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.210524 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-key podName:1f1b8f79-da9c-4a84-8d59-d8480d3f365e nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.710514116 +0000 UTC m=+141.617261618 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-key") pod "service-ca-9c57cc56f-td8vc" (UID: "1f1b8f79-da9c-4a84-8d59-d8480d3f365e") : failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.210497 4789 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.210541 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-srv-cert podName:4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9 nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.710534007 +0000 UTC m=+141.617281509 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-srv-cert") pod "catalog-operator-68c6474976-bkblm" (UID: "4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9") : failed to sync secret cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.210570 4789 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.210582 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-cabundle podName:1f1b8f79-da9c-4a84-8d59-d8480d3f365e nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.710554437 +0000 UTC m=+141.617301929 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-cabundle") pod "service-ca-9c57cc56f-td8vc" (UID: "1f1b8f79-da9c-4a84-8d59-d8480d3f365e") : failed to sync configmap cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: E1008 14:03:21.210603 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2389a066-e146-4ea8-bb92-d596311b564a-config-volume podName:2389a066-e146-4ea8-bb92-d596311b564a nodeName:}" failed. No retries permitted until 2025-10-08 14:03:21.710595419 +0000 UTC m=+141.617342921 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/2389a066-e146-4ea8-bb92-d596311b564a-config-volume") pod "collect-profiles-29332200-8zphp" (UID: "2389a066-e146-4ea8-bb92-d596311b564a") : failed to sync configmap cache: timed out waiting for the condition Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.215511 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.234735 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.256367 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.275437 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.295142 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.314625 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.335356 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.356008 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.375361 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.395170 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.415843 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.435493 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.455570 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.476010 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.496340 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.509445 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" event={"ID":"a9100571-6717-43fe-a0d8-0da2bd0ce31c","Type":"ContainerStarted","Data":"6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5"} Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.509493 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" event={"ID":"a9100571-6717-43fe-a0d8-0da2bd0ce31c","Type":"ContainerStarted","Data":"54be8b8f18e145300894a81e3624eb13ca8a917defdee2c30c3708705bd12c4f"} Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.509734 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.511471 4789 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rhnqk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.511519 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" podUID="a9100571-6717-43fe-a0d8-0da2bd0ce31c" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.515639 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.534825 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.554523 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.575264 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.595311 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.616387 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.635565 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.676916 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.695709 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.716149 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.731439 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-cabundle\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.731502 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-key\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.731657 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-srv-cert\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.731727 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-serving-cert\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.731786 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.731841 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-config\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.731878 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cbd6f891-627b-4f0b-ac4d-40e30d01fb89-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c8rg8\" (UID: \"cbd6f891-627b-4f0b-ac4d-40e30d01fb89\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.731962 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2389a066-e146-4ea8-bb92-d596311b564a-secret-volume\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.732084 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/73df2e86-6ba0-47d8-bfac-ef6d675808e6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hzztl\" (UID: \"73df2e86-6ba0-47d8-bfac-ef6d675808e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.732142 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2389a066-e146-4ea8-bb92-d596311b564a-config-volume\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.732945 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-cabundle\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.733214 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-config\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.733387 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2389a066-e146-4ea8-bb92-d596311b564a-config-volume\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.735412 4789 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.736424 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.737873 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cbd6f891-627b-4f0b-ac4d-40e30d01fb89-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-c8rg8\" (UID: \"cbd6f891-627b-4f0b-ac4d-40e30d01fb89\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.738487 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-serving-cert\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.739062 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/73df2e86-6ba0-47d8-bfac-ef6d675808e6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hzztl\" (UID: \"73df2e86-6ba0-47d8-bfac-ef6d675808e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.739826 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-srv-cert\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.740580 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-signing-key\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.741501 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2389a066-e146-4ea8-bb92-d596311b564a-secret-volume\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.755767 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.776146 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.821755 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwgzr\" (UniqueName: \"kubernetes.io/projected/b5199b69-0cb4-4958-abaf-41089b4f0559-kube-api-access-qwgzr\") pod \"console-operator-58897d9998-lkm5j\" (UID: \"b5199b69-0cb4-4958-abaf-41089b4f0559\") " pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.834515 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.835041 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.855856 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.874771 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.913295 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhxkb\" (UniqueName: \"kubernetes.io/projected/d58bece2-01c5-44eb-97ec-0de6757dd278-kube-api-access-hhxkb\") pod \"openshift-apiserver-operator-796bbdcf4f-lz74k\" (UID: \"d58bece2-01c5-44eb-97ec-0de6757dd278\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.931698 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg7l8\" (UniqueName: \"kubernetes.io/projected/119eb4b6-56d0-4cae-803a-56020cda0096-kube-api-access-gg7l8\") pod \"oauth-openshift-558db77b4-mtgx9\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.946480 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.954534 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh58c\" (UniqueName: \"kubernetes.io/projected/60295e4c-18c7-4b3d-acbf-54964ed76081-kube-api-access-rh58c\") pod \"machine-approver-56656f9798-vtftw\" (UID: \"60295e4c-18c7-4b3d-acbf-54964ed76081\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.978203 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hssvh\" (UniqueName: \"kubernetes.io/projected/6fb689b3-3b0d-476a-ba26-f6836061f366-kube-api-access-hssvh\") pod \"downloads-7954f5f757-mvzv5\" (UID: \"6fb689b3-3b0d-476a-ba26-f6836061f366\") " pod="openshift-console/downloads-7954f5f757-mvzv5" Oct 08 14:03:21 crc kubenswrapper[4789]: I1008 14:03:21.993947 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz749\" (UniqueName: \"kubernetes.io/projected/1764bb56-7b6e-4535-8af1-7d592163c9d1-kube-api-access-nz749\") pod \"authentication-operator-69f744f599-nbtdd\" (UID: \"1764bb56-7b6e-4535-8af1-7d592163c9d1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.015076 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmnn6\" (UniqueName: \"kubernetes.io/projected/5b2ca499-e6ad-4c98-9a03-fed4a44679c2-kube-api-access-xmnn6\") pod \"cluster-image-registry-operator-dc59b4c8b-ktcl9\" (UID: \"5b2ca499-e6ad-4c98-9a03-fed4a44679c2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.016653 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.036384 4789 request.go:700] Waited for 1.932275557s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.046848 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6dgm\" (UniqueName: \"kubernetes.io/projected/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-kube-api-access-l6dgm\") pod \"route-controller-manager-6576b87f9c-7gr6d\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.058595 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwjzz\" (UniqueName: \"kubernetes.io/projected/6fd06952-1b3e-44e6-af19-dfa31965b9ea-kube-api-access-zwjzz\") pod \"console-f9d7485db-dwjgm\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.074773 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk4kw\" (UniqueName: \"kubernetes.io/projected/bade0675-1189-4035-9673-2d20edbb729d-kube-api-access-hk4kw\") pod \"dns-operator-744455d44c-rmwkt\" (UID: \"bade0675-1189-4035-9673-2d20edbb729d\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.090703 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p8t9\" (UniqueName: \"kubernetes.io/projected/7191f22d-ffc0-47ff-b226-77f9213db94e-kube-api-access-4p8t9\") pod \"openshift-config-operator-7777fb866f-zmdbz\" (UID: \"7191f22d-ffc0-47ff-b226-77f9213db94e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.108755 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxld7\" (UniqueName: \"kubernetes.io/projected/6970b740-c77f-4817-b729-46839dc0698a-kube-api-access-vxld7\") pod \"apiserver-76f77b778f-kch4x\" (UID: \"6970b740-c77f-4817-b729-46839dc0698a\") " pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.133401 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.134812 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drd2q\" (UniqueName: \"kubernetes.io/projected/e7dc58a5-148e-45bb-9895-a43fe46a0064-kube-api-access-drd2q\") pod \"apiserver-7bbb656c7d-rh68r\" (UID: \"e7dc58a5-148e-45bb-9895-a43fe46a0064\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.151802 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.154339 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p78jx\" (UniqueName: \"kubernetes.io/projected/53dc6d10-02c4-48c5-8461-85ed095bcc11-kube-api-access-p78jx\") pod \"cluster-samples-operator-665b6dd947-7pwrl\" (UID: \"53dc6d10-02c4-48c5-8461-85ed095bcc11\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.155243 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lkm5j"] Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.156894 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.161513 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.168719 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.174951 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.176588 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.183580 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.193046 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mvzv5" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.194812 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.200568 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.216636 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.222696 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k"] Oct 08 14:03:22 crc kubenswrapper[4789]: W1008 14:03:22.241358 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd58bece2_01c5_44eb_97ec_0de6757dd278.slice/crio-e1ca69818f27a5a527ec8a9bd46f2e9a09d04fe5eef0bcc67f6b3e6c8411f0a7 WatchSource:0}: Error finding container e1ca69818f27a5a527ec8a9bd46f2e9a09d04fe5eef0bcc67f6b3e6c8411f0a7: Status 404 returned error can't find the container with id e1ca69818f27a5a527ec8a9bd46f2e9a09d04fe5eef0bcc67f6b3e6c8411f0a7 Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.261068 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg4pz\" (UniqueName: \"kubernetes.io/projected/bc2cc553-6ac7-44a6-948c-c1363836923d-kube-api-access-bg4pz\") pod \"machine-api-operator-5694c8668f-wmqxc\" (UID: \"bc2cc553-6ac7-44a6-948c-c1363836923d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.265730 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.268475 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.275855 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.293795 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7rt7\" (UniqueName: \"kubernetes.io/projected/73df2e86-6ba0-47d8-bfac-ef6d675808e6-kube-api-access-p7rt7\") pod \"package-server-manager-789f6589d5-hzztl\" (UID: \"73df2e86-6ba0-47d8-bfac-ef6d675808e6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.296777 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m7tl\" (UniqueName: \"kubernetes.io/projected/f8317867-1744-48e0-8940-35c3a0c2ab1b-kube-api-access-6m7tl\") pod \"openshift-controller-manager-operator-756b6f6bc6-g9tdd\" (UID: \"f8317867-1744-48e0-8940-35c3a0c2ab1b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.314323 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6kxj\" (UniqueName: \"kubernetes.io/projected/aa675ad1-38c1-49b7-9572-f165efdbd2e9-kube-api-access-m6kxj\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.334345 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnps4\" (UniqueName: \"kubernetes.io/projected/1f1b8f79-da9c-4a84-8d59-d8480d3f365e-kube-api-access-lnps4\") pod \"service-ca-9c57cc56f-td8vc\" (UID: \"1f1b8f79-da9c-4a84-8d59-d8480d3f365e\") " pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.361920 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv4wp\" (UniqueName: \"kubernetes.io/projected/1c546fd5-9ea3-402e-b837-dca47372c52c-kube-api-access-wv4wp\") pod \"machine-config-controller-84d6567774-rbvwr\" (UID: \"1c546fd5-9ea3-402e-b837-dca47372c52c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.373330 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.377618 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6jt8\" (UniqueName: \"kubernetes.io/projected/172550ec-f5a7-4fb2-8c89-8ec82e2775f7-kube-api-access-r6jt8\") pod \"packageserver-d55dfcdfc-84p6z\" (UID: \"172550ec-f5a7-4fb2-8c89-8ec82e2775f7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.391516 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.392308 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.402493 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/449f550c-38c6-4c62-abc7-2566c2cf1ccc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l7dth\" (UID: \"449f550c-38c6-4c62-abc7-2566c2cf1ccc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.409664 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7jpl\" (UniqueName: \"kubernetes.io/projected/8e83dccc-9fa2-42d5-ae9a-7941ae00d321-kube-api-access-x7jpl\") pod \"service-ca-operator-777779d784-kx5ls\" (UID: \"8e83dccc-9fa2-42d5-ae9a-7941ae00d321\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.418582 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.431118 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b9ht\" (UniqueName: \"kubernetes.io/projected/cbd6f891-627b-4f0b-ac4d-40e30d01fb89-kube-api-access-9b9ht\") pod \"control-plane-machine-set-operator-78cbb6b69f-c8rg8\" (UID: \"cbd6f891-627b-4f0b-ac4d-40e30d01fb89\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.436244 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.449816 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlj85\" (UniqueName: \"kubernetes.io/projected/09101a18-6c60-4da9-b517-afb6ae89cb4c-kube-api-access-nlj85\") pod \"router-default-5444994796-p2mjq\" (UID: \"09101a18-6c60-4da9-b517-afb6ae89cb4c\") " pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.456733 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.477828 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h9lt\" (UniqueName: \"kubernetes.io/projected/5b3f76e2-8cce-47f1-87e5-43be0d07238e-kube-api-access-2h9lt\") pod \"marketplace-operator-79b997595-vzq6g\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.494495 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r67d6\" (UniqueName: \"kubernetes.io/projected/2389a066-e146-4ea8-bb92-d596311b564a-kube-api-access-r67d6\") pod \"collect-profiles-29332200-8zphp\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.513454 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" event={"ID":"d58bece2-01c5-44eb-97ec-0de6757dd278","Type":"ContainerStarted","Data":"e1ca69818f27a5a527ec8a9bd46f2e9a09d04fe5eef0bcc67f6b3e6c8411f0a7"} Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.515016 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lkm5j" event={"ID":"b5199b69-0cb4-4958-abaf-41089b4f0559","Type":"ContainerStarted","Data":"f185df3f3bf287ee6e256bdc94a57f4c6bab274fc80672104a6572cff1b91867"} Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.515252 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wvkn\" (UniqueName: \"kubernetes.io/projected/9d8a129c-9c83-4ed2-a9c5-b017a0bd470b-kube-api-access-2wvkn\") pod \"machine-config-operator-74547568cd-xpczn\" (UID: \"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.517510 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" event={"ID":"60295e4c-18c7-4b3d-acbf-54964ed76081","Type":"ContainerStarted","Data":"ee195dba388e85d2d1b339566ddb1fd30d1f9356497b00bf0aa8a70dbac3d435"} Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.518036 4789 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rhnqk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.518067 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" podUID="a9100571-6717-43fe-a0d8-0da2bd0ce31c" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.530556 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6psjr\" (UniqueName: \"kubernetes.io/projected/17ebe9cd-3f39-42c3-a888-7270e67caa06-kube-api-access-6psjr\") pod \"etcd-operator-b45778765-hf8hk\" (UID: \"17ebe9cd-3f39-42c3-a888-7270e67caa06\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.556758 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6ae8586c-f105-42ff-b845-b4eb6a878694-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dqrvj\" (UID: \"6ae8586c-f105-42ff-b845-b4eb6a878694\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.561980 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.573714 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crwnl\" (UniqueName: \"kubernetes.io/projected/4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9-kube-api-access-crwnl\") pod \"catalog-operator-68c6474976-bkblm\" (UID: \"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.595322 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.599305 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa675ad1-38c1-49b7-9572-f165efdbd2e9-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qlf4j\" (UID: \"aa675ad1-38c1-49b7-9572-f165efdbd2e9\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.601766 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.609750 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.631845 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.640009 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kch4x"] Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.648597 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d"] Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.650165 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651193 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651242 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-trusted-ca\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651279 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc66affb-e030-498d-819c-58646b279757-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-prj2m\" (UID: \"dc66affb-e030-498d-819c-58646b279757\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651306 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5a5f8e9f-467a-4ec3-912a-24dbee088302-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xs8k5\" (UID: \"5a5f8e9f-467a-4ec3-912a-24dbee088302\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651558 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-tls\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: E1008 14:03:22.651620 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:23.151598146 +0000 UTC m=+143.058345638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651664 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651706 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651735 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvvws\" (UniqueName: \"kubernetes.io/projected/7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6-kube-api-access-nvvws\") pod \"olm-operator-6b444d44fb-k4wrp\" (UID: \"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651788 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-drjkk\" (UID: \"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651811 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-bound-sa-token\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651858 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-drjkk\" (UID: \"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651875 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6-srv-cert\") pod \"olm-operator-6b444d44fb-k4wrp\" (UID: \"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651896 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-certificates\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651930 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-888bg\" (UniqueName: \"kubernetes.io/projected/d79e33d9-6682-4659-8450-bfcca4c46eb6-kube-api-access-888bg\") pod \"migrator-59844c95c7-ljh2s\" (UID: \"d79e33d9-6682-4659-8450-bfcca4c46eb6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651951 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf9qx\" (UniqueName: \"kubernetes.io/projected/26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6-kube-api-access-bf9qx\") pod \"kube-storage-version-migrator-operator-b67b599dd-drjkk\" (UID: \"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.651979 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw9dh\" (UniqueName: \"kubernetes.io/projected/5a5f8e9f-467a-4ec3-912a-24dbee088302-kube-api-access-kw9dh\") pod \"multus-admission-controller-857f4d67dd-xs8k5\" (UID: \"5a5f8e9f-467a-4ec3-912a-24dbee088302\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.652063 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q2mw\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-kube-api-access-8q2mw\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.652088 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k4wrp\" (UID: \"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.652165 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc66affb-e030-498d-819c-58646b279757-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-prj2m\" (UID: \"dc66affb-e030-498d-819c-58646b279757\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.652207 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc66affb-e030-498d-819c-58646b279757-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-prj2m\" (UID: \"dc66affb-e030-498d-819c-58646b279757\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.659308 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.668480 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.702547 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.728383 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.746269 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.753224 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:22 crc kubenswrapper[4789]: E1008 14:03:22.753495 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:23.253459331 +0000 UTC m=+143.160206823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.753593 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc66affb-e030-498d-819c-58646b279757-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-prj2m\" (UID: \"dc66affb-e030-498d-819c-58646b279757\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.753652 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5a5f8e9f-467a-4ec3-912a-24dbee088302-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xs8k5\" (UID: \"5a5f8e9f-467a-4ec3-912a-24dbee088302\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.753710 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vk2l\" (UniqueName: \"kubernetes.io/projected/df648835-daa7-48df-aa95-0d56ad850f0b-kube-api-access-7vk2l\") pod \"machine-config-server-jgbjw\" (UID: \"df648835-daa7-48df-aa95-0d56ad850f0b\") " pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.753764 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-tls\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.753789 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cf5270f-6d48-40ad-9332-118daa09f473-config-volume\") pod \"dns-default-gr58t\" (UID: \"8cf5270f-6d48-40ad-9332-118daa09f473\") " pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.753848 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.753907 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.753964 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-plugins-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754001 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8cf5270f-6d48-40ad-9332-118daa09f473-metrics-tls\") pod \"dns-default-gr58t\" (UID: \"8cf5270f-6d48-40ad-9332-118daa09f473\") " pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754048 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvvws\" (UniqueName: \"kubernetes.io/projected/7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6-kube-api-access-nvvws\") pod \"olm-operator-6b444d44fb-k4wrp\" (UID: \"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754064 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-registration-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754091 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-socket-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754118 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kczfh\" (UniqueName: \"kubernetes.io/projected/2b2efb14-dde5-4a97-9661-9ff6ec231414-kube-api-access-kczfh\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754215 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-drjkk\" (UID: \"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754249 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwq5m\" (UniqueName: \"kubernetes.io/projected/8cf5270f-6d48-40ad-9332-118daa09f473-kube-api-access-zwq5m\") pod \"dns-default-gr58t\" (UID: \"8cf5270f-6d48-40ad-9332-118daa09f473\") " pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754299 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-bound-sa-token\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754441 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-csi-data-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754483 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-mountpoint-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754554 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsmq7\" (UniqueName: \"kubernetes.io/projected/32a50c76-0675-4ff6-b8d2-0239416ed184-kube-api-access-hsmq7\") pod \"ingress-canary-dncbh\" (UID: \"32a50c76-0675-4ff6-b8d2-0239416ed184\") " pod="openshift-ingress-canary/ingress-canary-dncbh" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754594 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-drjkk\" (UID: \"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754611 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6-srv-cert\") pod \"olm-operator-6b444d44fb-k4wrp\" (UID: \"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754628 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/df648835-daa7-48df-aa95-0d56ad850f0b-certs\") pod \"machine-config-server-jgbjw\" (UID: \"df648835-daa7-48df-aa95-0d56ad850f0b\") " pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754678 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-certificates\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754836 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754955 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-888bg\" (UniqueName: \"kubernetes.io/projected/d79e33d9-6682-4659-8450-bfcca4c46eb6-kube-api-access-888bg\") pod \"migrator-59844c95c7-ljh2s\" (UID: \"d79e33d9-6682-4659-8450-bfcca4c46eb6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.754998 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/df648835-daa7-48df-aa95-0d56ad850f0b-node-bootstrap-token\") pod \"machine-config-server-jgbjw\" (UID: \"df648835-daa7-48df-aa95-0d56ad850f0b\") " pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.755052 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf9qx\" (UniqueName: \"kubernetes.io/projected/26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6-kube-api-access-bf9qx\") pod \"kube-storage-version-migrator-operator-b67b599dd-drjkk\" (UID: \"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.755169 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw9dh\" (UniqueName: \"kubernetes.io/projected/5a5f8e9f-467a-4ec3-912a-24dbee088302-kube-api-access-kw9dh\") pod \"multus-admission-controller-857f4d67dd-xs8k5\" (UID: \"5a5f8e9f-467a-4ec3-912a-24dbee088302\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.755232 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/32a50c76-0675-4ff6-b8d2-0239416ed184-cert\") pod \"ingress-canary-dncbh\" (UID: \"32a50c76-0675-4ff6-b8d2-0239416ed184\") " pod="openshift-ingress-canary/ingress-canary-dncbh" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.755327 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q2mw\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-kube-api-access-8q2mw\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.755367 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k4wrp\" (UID: \"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.755412 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc66affb-e030-498d-819c-58646b279757-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-prj2m\" (UID: \"dc66affb-e030-498d-819c-58646b279757\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.755447 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc66affb-e030-498d-819c-58646b279757-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-prj2m\" (UID: \"dc66affb-e030-498d-819c-58646b279757\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.758381 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.758725 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-trusted-ca\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.760781 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc66affb-e030-498d-819c-58646b279757-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-prj2m\" (UID: \"dc66affb-e030-498d-819c-58646b279757\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:22 crc kubenswrapper[4789]: E1008 14:03:22.765315 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:23.265293097 +0000 UTC m=+143.172040579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.766315 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.767374 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5a5f8e9f-467a-4ec3-912a-24dbee088302-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-xs8k5\" (UID: \"5a5f8e9f-467a-4ec3-912a-24dbee088302\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.769604 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-drjkk\" (UID: \"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.769832 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc66affb-e030-498d-819c-58646b279757-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-prj2m\" (UID: \"dc66affb-e030-498d-819c-58646b279757\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.782257 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-drjkk\" (UID: \"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.787253 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-tls\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.788853 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k4wrp\" (UID: \"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.789656 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvvws\" (UniqueName: \"kubernetes.io/projected/7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6-kube-api-access-nvvws\") pod \"olm-operator-6b444d44fb-k4wrp\" (UID: \"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.789770 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6-srv-cert\") pod \"olm-operator-6b444d44fb-k4wrp\" (UID: \"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.811842 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-888bg\" (UniqueName: \"kubernetes.io/projected/d79e33d9-6682-4659-8450-bfcca4c46eb6-kube-api-access-888bg\") pod \"migrator-59844c95c7-ljh2s\" (UID: \"d79e33d9-6682-4659-8450-bfcca4c46eb6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.817081 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz"] Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.828730 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-trusted-ca\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.829724 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-certificates\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.856041 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf9qx\" (UniqueName: \"kubernetes.io/projected/26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6-kube-api-access-bf9qx\") pod \"kube-storage-version-migrator-operator-b67b599dd-drjkk\" (UID: \"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887133 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887348 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vk2l\" (UniqueName: \"kubernetes.io/projected/df648835-daa7-48df-aa95-0d56ad850f0b-kube-api-access-7vk2l\") pod \"machine-config-server-jgbjw\" (UID: \"df648835-daa7-48df-aa95-0d56ad850f0b\") " pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887375 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cf5270f-6d48-40ad-9332-118daa09f473-config-volume\") pod \"dns-default-gr58t\" (UID: \"8cf5270f-6d48-40ad-9332-118daa09f473\") " pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887407 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-plugins-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887428 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8cf5270f-6d48-40ad-9332-118daa09f473-metrics-tls\") pod \"dns-default-gr58t\" (UID: \"8cf5270f-6d48-40ad-9332-118daa09f473\") " pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887451 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-socket-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887473 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-registration-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887496 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kczfh\" (UniqueName: \"kubernetes.io/projected/2b2efb14-dde5-4a97-9661-9ff6ec231414-kube-api-access-kczfh\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887536 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwq5m\" (UniqueName: \"kubernetes.io/projected/8cf5270f-6d48-40ad-9332-118daa09f473-kube-api-access-zwq5m\") pod \"dns-default-gr58t\" (UID: \"8cf5270f-6d48-40ad-9332-118daa09f473\") " pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887566 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-csi-data-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887592 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-mountpoint-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887616 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsmq7\" (UniqueName: \"kubernetes.io/projected/32a50c76-0675-4ff6-b8d2-0239416ed184-kube-api-access-hsmq7\") pod \"ingress-canary-dncbh\" (UID: \"32a50c76-0675-4ff6-b8d2-0239416ed184\") " pod="openshift-ingress-canary/ingress-canary-dncbh" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887636 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/df648835-daa7-48df-aa95-0d56ad850f0b-certs\") pod \"machine-config-server-jgbjw\" (UID: \"df648835-daa7-48df-aa95-0d56ad850f0b\") " pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887665 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/df648835-daa7-48df-aa95-0d56ad850f0b-node-bootstrap-token\") pod \"machine-config-server-jgbjw\" (UID: \"df648835-daa7-48df-aa95-0d56ad850f0b\") " pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887696 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/32a50c76-0675-4ff6-b8d2-0239416ed184-cert\") pod \"ingress-canary-dncbh\" (UID: \"32a50c76-0675-4ff6-b8d2-0239416ed184\") " pod="openshift-ingress-canary/ingress-canary-dncbh" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.887856 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-plugins-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: E1008 14:03:22.887963 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:23.387937035 +0000 UTC m=+143.294684527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.888731 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cf5270f-6d48-40ad-9332-118daa09f473-config-volume\") pod \"dns-default-gr58t\" (UID: \"8cf5270f-6d48-40ad-9332-118daa09f473\") " pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.890857 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-csi-data-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.890941 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-mountpoint-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.891048 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-registration-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.891109 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2b2efb14-dde5-4a97-9661-9ff6ec231414-socket-dir\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.895472 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/df648835-daa7-48df-aa95-0d56ad850f0b-certs\") pod \"machine-config-server-jgbjw\" (UID: \"df648835-daa7-48df-aa95-0d56ad850f0b\") " pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.896272 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.902170 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/df648835-daa7-48df-aa95-0d56ad850f0b-node-bootstrap-token\") pod \"machine-config-server-jgbjw\" (UID: \"df648835-daa7-48df-aa95-0d56ad850f0b\") " pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.903289 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/32a50c76-0675-4ff6-b8d2-0239416ed184-cert\") pod \"ingress-canary-dncbh\" (UID: \"32a50c76-0675-4ff6-b8d2-0239416ed184\") " pod="openshift-ingress-canary/ingress-canary-dncbh" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.904113 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nbtdd"] Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.904242 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8cf5270f-6d48-40ad-9332-118daa09f473-metrics-tls\") pod \"dns-default-gr58t\" (UID: \"8cf5270f-6d48-40ad-9332-118daa09f473\") " pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.904775 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw9dh\" (UniqueName: \"kubernetes.io/projected/5a5f8e9f-467a-4ec3-912a-24dbee088302-kube-api-access-kw9dh\") pod \"multus-admission-controller-857f4d67dd-xs8k5\" (UID: \"5a5f8e9f-467a-4ec3-912a-24dbee088302\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.905470 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mvzv5"] Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.906835 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-bound-sa-token\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.915260 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q2mw\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-kube-api-access-8q2mw\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.932765 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc66affb-e030-498d-819c-58646b279757-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-prj2m\" (UID: \"dc66affb-e030-498d-819c-58646b279757\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.940449 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mtgx9"] Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.975705 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.986492 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.988869 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:22 crc kubenswrapper[4789]: E1008 14:03:22.989214 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:23.489201144 +0000 UTC m=+143.395948636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.995493 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vk2l\" (UniqueName: \"kubernetes.io/projected/df648835-daa7-48df-aa95-0d56ad850f0b-kube-api-access-7vk2l\") pod \"machine-config-server-jgbjw\" (UID: \"df648835-daa7-48df-aa95-0d56ad850f0b\") " pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:22 crc kubenswrapper[4789]: I1008 14:03:22.997420 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwq5m\" (UniqueName: \"kubernetes.io/projected/8cf5270f-6d48-40ad-9332-118daa09f473-kube-api-access-zwq5m\") pod \"dns-default-gr58t\" (UID: \"8cf5270f-6d48-40ad-9332-118daa09f473\") " pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.013450 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.018053 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsmq7\" (UniqueName: \"kubernetes.io/projected/32a50c76-0675-4ff6-b8d2-0239416ed184-kube-api-access-hsmq7\") pod \"ingress-canary-dncbh\" (UID: \"32a50c76-0675-4ff6-b8d2-0239416ed184\") " pod="openshift-ingress-canary/ingress-canary-dncbh" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.041370 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kczfh\" (UniqueName: \"kubernetes.io/projected/2b2efb14-dde5-4a97-9661-9ff6ec231414-kube-api-access-kczfh\") pod \"csi-hostpathplugin-96vg6\" (UID: \"2b2efb14-dde5-4a97-9661-9ff6ec231414\") " pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.065788 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.073463 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-jgbjw" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.091503 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.091942 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:23.591917743 +0000 UTC m=+143.498665235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.102108 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-96vg6" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.109370 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.116899 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dncbh" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.169646 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" podStartSLOduration=122.169628953 podStartE2EDuration="2m2.169628953s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:23.168319317 +0000 UTC m=+143.075066809" watchObservedRunningTime="2025-10-08 14:03:23.169628953 +0000 UTC m=+143.076376445" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.173653 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-dwjgm"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.175843 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rmwkt"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.187377 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.196206 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.196569 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:23.696556655 +0000 UTC m=+143.603304147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.239881 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wmqxc"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.289677 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.297356 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.297669 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:23.797650499 +0000 UTC m=+143.704397981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.318115 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.319845 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.360313 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.390002 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-td8vc"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.406525 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.407611 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.409612 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:23.909593212 +0000 UTC m=+143.816340704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.441296 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.509385 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.509843 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.009824812 +0000 UTC m=+143.916572304 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.510007 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.510348 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.010341397 +0000 UTC m=+143.917088879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.557498 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.566934 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" event={"ID":"1f1b8f79-da9c-4a84-8d59-d8480d3f365e","Type":"ContainerStarted","Data":"91dcc7ca720c412ea990b32de3575b53eaf240dc8fdacdca122c5669d2c04afb"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.567947 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.568933 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hf8hk"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.570090 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" event={"ID":"60295e4c-18c7-4b3d-acbf-54964ed76081","Type":"ContainerStarted","Data":"2390ab5d55df1a4b635178f71c752922d52c3f9344b66224b0dc245dac38836d"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.571226 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" event={"ID":"5ae1d9aa-857d-416d-92a1-e08e45bd7b19","Type":"ContainerStarted","Data":"61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.571249 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" event={"ID":"5ae1d9aa-857d-416d-92a1-e08e45bd7b19","Type":"ContainerStarted","Data":"52a12cd8b8724a5076c9d4e6809ba6c6d2ada7c85022c0b173dc5f309d74fe4d"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.573060 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.574530 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dwjgm" event={"ID":"6fd06952-1b3e-44e6-af19-dfa31965b9ea","Type":"ContainerStarted","Data":"7b1a261d86306329063009babd8ea59e76f366a02454369d3eb7b04f06126653"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.575452 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" event={"ID":"1764bb56-7b6e-4535-8af1-7d592163c9d1","Type":"ContainerStarted","Data":"b785e19aef2696edc782840b2dbd0710bfe2fed778df9e057bba779f221401ef"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.576636 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" event={"ID":"73df2e86-6ba0-47d8-bfac-ef6d675808e6","Type":"ContainerStarted","Data":"318cbf281b55c87ff36ccc7ce53de975a4cdaf27fd449fb70725920cfe71349e"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.577898 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" event={"ID":"bc2cc553-6ac7-44a6-948c-c1363836923d","Type":"ContainerStarted","Data":"f00f71fa37939a08054ed02f941e934242b4bd18012c0c1292f6dea1e1de0343"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.581917 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" event={"ID":"6970b740-c77f-4817-b729-46839dc0698a","Type":"ContainerStarted","Data":"f0e7a882a38cf632290c18e313a8edfe6720ae0c1a725bd553de4bc1f822d14d"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.583283 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" event={"ID":"cbd6f891-627b-4f0b-ac4d-40e30d01fb89","Type":"ContainerStarted","Data":"20bc5995dacd4b7ffe984f81e808a93822144408b0715320edd3a3c923ff4709"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.588556 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" event={"ID":"bade0675-1189-4035-9673-2d20edbb729d","Type":"ContainerStarted","Data":"25a5f2a2311326f3878c748b74364e56d08de55c6148af7cfeb0f24e9915503d"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.596435 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" event={"ID":"5b2ca499-e6ad-4c98-9a03-fed4a44679c2","Type":"ContainerStarted","Data":"6abbe4b5128cf98064fc6845a4322659762a6d6ad91d1b9764cf8e9fc7877a57"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.597259 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-jgbjw" event={"ID":"df648835-daa7-48df-aa95-0d56ad850f0b","Type":"ContainerStarted","Data":"17f7b38bc649c09aa3af0ac4982de3c19cadde2ea5ca8f611dc793e355b5de34"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.597826 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" event={"ID":"119eb4b6-56d0-4cae-803a-56020cda0096","Type":"ContainerStarted","Data":"ad085a17b1cdcc909bab079580c0bb0485e954897be3aeb2ed0f0f126ed44603"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.598945 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lkm5j" event={"ID":"b5199b69-0cb4-4958-abaf-41089b4f0559","Type":"ContainerStarted","Data":"39e18f5c69aea91cfb4388da32044d6e14ac66722f0fb8a7ed8db0786f031fdc"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.599433 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.600448 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" event={"ID":"7191f22d-ffc0-47ff-b226-77f9213db94e","Type":"ContainerStarted","Data":"9e5838d5a6374d76a2b53c7184da365e79ba2d3b8e78bd655088928d118d206b"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.603091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" event={"ID":"d58bece2-01c5-44eb-97ec-0de6757dd278","Type":"ContainerStarted","Data":"2d44e8e89af11e40607b16d2fd194e01e26b88cc05872075ec3a9d5812fefd44"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.604853 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p2mjq" event={"ID":"09101a18-6c60-4da9-b517-afb6ae89cb4c","Type":"ContainerStarted","Data":"c5c73caa0f60247202a0f7dc403489e319daf2f2001cea07f1cb7c295b813bd1"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.605631 4789 patch_prober.go:28] interesting pod/console-operator-58897d9998-lkm5j container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.605669 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-lkm5j" podUID="b5199b69-0cb4-4958-abaf-41089b4f0559" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.605947 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" event={"ID":"172550ec-f5a7-4fb2-8c89-8ec82e2775f7","Type":"ContainerStarted","Data":"cf9de03b5df707e6c0fcd085129faed0922c6a4ed1637507655fb4d1d404210a"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.615596 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.615923 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.115906864 +0000 UTC m=+144.022654356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.624687 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mvzv5" event={"ID":"6fb689b3-3b0d-476a-ba26-f6836061f366","Type":"ContainerStarted","Data":"1397d8e860fd95e5b484b40aa1c8edc22114f7a8848c20aabdc743056dfce862"} Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.665700 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.680472 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.684134 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.694223 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.697347 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.698665 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vzq6g"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.707018 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.717598 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.718519 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.218505519 +0000 UTC m=+144.125253011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.820701 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.821045 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.321024612 +0000 UTC m=+144.227772104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.865131 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-xs8k5"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.877604 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.880836 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.885130 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk"] Oct 08 14:03:23 crc kubenswrapper[4789]: W1008 14:03:23.885658 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c546fd5_9ea3_402e_b837_dca47372c52c.slice/crio-4fdb141f8cc8de37c44ba35e1b3d40a03a84697bdf3c61773568130727de7e15 WatchSource:0}: Error finding container 4fdb141f8cc8de37c44ba35e1b3d40a03a84697bdf3c61773568130727de7e15: Status 404 returned error can't find the container with id 4fdb141f8cc8de37c44ba35e1b3d40a03a84697bdf3c61773568130727de7e15 Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.920280 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dncbh"] Oct 08 14:03:23 crc kubenswrapper[4789]: I1008 14:03:23.922082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:23 crc kubenswrapper[4789]: E1008 14:03:23.922440 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.422425765 +0000 UTC m=+144.329173257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: W1008 14:03:24.006085 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26ee44ad_a6fe_4a2c_bc67_ab856b48a0f6.slice/crio-814c7addca6a184648176d28a2b8ca4e60a8a8fec04d4585827e87ed98b9cd5a WatchSource:0}: Error finding container 814c7addca6a184648176d28a2b8ca4e60a8a8fec04d4585827e87ed98b9cd5a: Status 404 returned error can't find the container with id 814c7addca6a184648176d28a2b8ca4e60a8a8fec04d4585827e87ed98b9cd5a Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.008635 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m"] Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.031804 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gr58t"] Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.051442 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.051734 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.551714516 +0000 UTC m=+144.458462008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.053400 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-96vg6"] Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.154846 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.155375 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.65536093 +0000 UTC m=+144.562108412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.256207 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.256356 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.756313961 +0000 UTC m=+144.663061453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.256423 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.256852 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.756812894 +0000 UTC m=+144.663560386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.358496 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.359328 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.859306197 +0000 UTC m=+144.766053689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.460156 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.460568 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:24.960553536 +0000 UTC m=+144.867301028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.561267 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.561447 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.061412323 +0000 UTC m=+144.968159815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.561503 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.561946 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.061939828 +0000 UTC m=+144.968687320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.575931 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lz74k" podStartSLOduration=124.575915643 podStartE2EDuration="2m4.575915643s" podCreationTimestamp="2025-10-08 14:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:24.574676309 +0000 UTC m=+144.481423801" watchObservedRunningTime="2025-10-08 14:03:24.575915643 +0000 UTC m=+144.482663135" Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.663811 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.664203 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.164178604 +0000 UTC m=+145.070926096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.668128 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" event={"ID":"dc66affb-e030-498d-819c-58646b279757","Type":"ContainerStarted","Data":"58b3bdd9bd7b50001d844c290ca3d0a33248a752f18224cd31dad1e93046e4f0"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.670443 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" event={"ID":"5b2ca499-e6ad-4c98-9a03-fed4a44679c2","Type":"ContainerStarted","Data":"0eed0a22b30f34bc510d00fe6599a8245bd7b7663ccb6da874342dce453d4ce4"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.685118 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" event={"ID":"cbd6f891-627b-4f0b-ac4d-40e30d01fb89","Type":"ContainerStarted","Data":"f1a240d9f390dc2838455f684194ff742d0792d11eb702b23235068f81029a0d"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.694729 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-lkm5j" podStartSLOduration=123.694710575 podStartE2EDuration="2m3.694710575s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:24.616475 +0000 UTC m=+144.523222492" watchObservedRunningTime="2025-10-08 14:03:24.694710575 +0000 UTC m=+144.601458067" Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.704042 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96vg6" event={"ID":"2b2efb14-dde5-4a97-9661-9ff6ec231414","Type":"ContainerStarted","Data":"6985729dd3a6b7ffbbc35d2efed5da6aa75b45917ea29eb4181f0558b19a1f29"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.714487 4789 generic.go:334] "Generic (PLEG): container finished" podID="7191f22d-ffc0-47ff-b226-77f9213db94e" containerID="25ff563ec56249b26bb2a3c85836cd272ca3f09a8b18c6e0791b98a46b634a18" exitCode=0 Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.714783 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" event={"ID":"7191f22d-ffc0-47ff-b226-77f9213db94e","Type":"ContainerDied","Data":"25ff563ec56249b26bb2a3c85836cd272ca3f09a8b18c6e0791b98a46b634a18"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.716913 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ktcl9" podStartSLOduration=123.716875925 podStartE2EDuration="2m3.716875925s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:24.696208966 +0000 UTC m=+144.602956458" watchObservedRunningTime="2025-10-08 14:03:24.716875925 +0000 UTC m=+144.623623417" Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.718589 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-c8rg8" podStartSLOduration=123.718580332 podStartE2EDuration="2m3.718580332s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:24.7145118 +0000 UTC m=+144.621259302" watchObservedRunningTime="2025-10-08 14:03:24.718580332 +0000 UTC m=+144.625327824" Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.765500 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.767709 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.267660084 +0000 UTC m=+145.174407656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.773782 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" event={"ID":"17ebe9cd-3f39-42c3-a888-7270e67caa06","Type":"ContainerStarted","Data":"615d40d89d01809ae62d01059f8008616910a70896937d20dc239c89994f7598"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.774156 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" event={"ID":"1c546fd5-9ea3-402e-b837-dca47372c52c","Type":"ContainerStarted","Data":"6e923165f4df97d0a0a9a62caf0d0b0665b9bfe3e842eb0b0a58396a959f5bbb"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.774172 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" event={"ID":"1c546fd5-9ea3-402e-b837-dca47372c52c","Type":"ContainerStarted","Data":"4fdb141f8cc8de37c44ba35e1b3d40a03a84697bdf3c61773568130727de7e15"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.844674 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" event={"ID":"2389a066-e146-4ea8-bb92-d596311b564a","Type":"ContainerStarted","Data":"3b44d2293810e705c7b1f5e97e9bced29d6a0df57d91af766ea28f9a09936bf6"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.849650 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" event={"ID":"aa675ad1-38c1-49b7-9572-f165efdbd2e9","Type":"ContainerStarted","Data":"298a243370a789e5c04e530ec89e968c77418059d998206fc13572bb85bfa10d"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.852200 4789 generic.go:334] "Generic (PLEG): container finished" podID="6970b740-c77f-4817-b729-46839dc0698a" containerID="b3830c22375913a0549b431b3f4acfc9a90d8dd03f5d8abe6affeec7c27a53f4" exitCode=0 Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.852250 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" event={"ID":"6970b740-c77f-4817-b729-46839dc0698a","Type":"ContainerDied","Data":"b3830c22375913a0549b431b3f4acfc9a90d8dd03f5d8abe6affeec7c27a53f4"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.859787 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" event={"ID":"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9","Type":"ContainerStarted","Data":"7b1b850f990638102b51e959f76ee53b5078ea6a43d66815c4df7a793efeb45c"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.870715 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.871969 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.371947416 +0000 UTC m=+145.278694908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.930089 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" event={"ID":"6ae8586c-f105-42ff-b845-b4eb6a878694","Type":"ContainerStarted","Data":"9e48c7977fea07f4df18c41b59ab9ae188d21c053e86b94a073efa91ac2ff58c"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.937179 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" event={"ID":"bade0675-1189-4035-9673-2d20edbb729d","Type":"ContainerStarted","Data":"7982c27375a9d32729cb79ca58721b9b502528748e50d23e5fd9e66dae133cf7"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.945224 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" event={"ID":"1764bb56-7b6e-4535-8af1-7d592163c9d1","Type":"ContainerStarted","Data":"27ca4c1bf5dbf129f6dc2a2140663b52a02a8834a2fd788115efe793269f9add"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.953492 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" podStartSLOduration=124.953467671 podStartE2EDuration="2m4.953467671s" podCreationTimestamp="2025-10-08 14:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:24.875514214 +0000 UTC m=+144.782261706" watchObservedRunningTime="2025-10-08 14:03:24.953467671 +0000 UTC m=+144.860215163" Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.972818 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" event={"ID":"bc2cc553-6ac7-44a6-948c-c1363836923d","Type":"ContainerStarted","Data":"6c61c43378da208113706bc26df995e53aaadc34cf083f51bd020268730fbf91"} Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.973632 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:24 crc kubenswrapper[4789]: E1008 14:03:24.975492 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.475473457 +0000 UTC m=+145.382220949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:24 crc kubenswrapper[4789]: I1008 14:03:24.979590 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-nbtdd" podStartSLOduration=124.97957589 podStartE2EDuration="2m4.97957589s" podCreationTimestamp="2025-10-08 14:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:24.978323036 +0000 UTC m=+144.885070528" watchObservedRunningTime="2025-10-08 14:03:24.97957589 +0000 UTC m=+144.886323382" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.011298 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-jgbjw" event={"ID":"df648835-daa7-48df-aa95-0d56ad850f0b","Type":"ContainerStarted","Data":"bba5482a7e2c0d4d28e6caf325e081a399b07210073eaacc0d2549a0429ab5b4"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.013749 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" event={"ID":"f8317867-1744-48e0-8940-35c3a0c2ab1b","Type":"ContainerStarted","Data":"5d143a21e462a22e31565c1089851f5862b68aaef4d904be59b8167921d9a172"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.046800 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-jgbjw" podStartSLOduration=6.046779721 podStartE2EDuration="6.046779721s" podCreationTimestamp="2025-10-08 14:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:25.039268904 +0000 UTC m=+144.946016386" watchObservedRunningTime="2025-10-08 14:03:25.046779721 +0000 UTC m=+144.953527213" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.047605 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" event={"ID":"172550ec-f5a7-4fb2-8c89-8ec82e2775f7","Type":"ContainerStarted","Data":"1a59da0f1a78a71634279a55db16a79a744f611439eada4674869de75ef66de5"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.048343 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.062366 4789 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-84p6z container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.062425 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" podUID="172550ec-f5a7-4fb2-8c89-8ec82e2775f7" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.076040 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.081927 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" event={"ID":"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b","Type":"ContainerStarted","Data":"1c36afb8fbdf70106b3a6607b59fa8bc67423121f148bc6275f1ded4582e3bc1"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.081965 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" event={"ID":"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b","Type":"ContainerStarted","Data":"7bbdc311a4a12429cf703a403e4e09791f9cc83fa7a01c2e657fa9a7f840e983"} Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.083300 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.583276316 +0000 UTC m=+145.490023808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.099253 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" event={"ID":"449f550c-38c6-4c62-abc7-2566c2cf1ccc","Type":"ContainerStarted","Data":"f2b3a7410fb1fde9cf6501920806a35ea17e5ce0a6f8ab8171fe6e0ff6ebb38e"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.108259 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" event={"ID":"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6","Type":"ContainerStarted","Data":"1aa5236c924854b2a3e3d1890f47b70e7779d7d00110cf78c5d70f5dc0b3cac6"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.114589 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gr58t" event={"ID":"8cf5270f-6d48-40ad-9332-118daa09f473","Type":"ContainerStarted","Data":"c540b79b2b2848cc0cc74b1f3f730f964a51d62282ca1642eb2ed0e42777d49c"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.147020 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dncbh" event={"ID":"32a50c76-0675-4ff6-b8d2-0239416ed184","Type":"ContainerStarted","Data":"11b9787b489b83e4f2135c7227398e5331f3a4fb990b877037eca91ab0e72d95"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.157728 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" event={"ID":"119eb4b6-56d0-4cae-803a-56020cda0096","Type":"ContainerStarted","Data":"9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.158116 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.178335 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.179144 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.679113336 +0000 UTC m=+145.585860818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.181156 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" event={"ID":"5b3f76e2-8cce-47f1-87e5-43be0d07238e","Type":"ContainerStarted","Data":"d37c36b6c06c86b617663fcacc4e1da5d66a2a20d62bee87bac3d374351ec4a0"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.190171 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.195590 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" podStartSLOduration=125.195563279 podStartE2EDuration="2m5.195563279s" podCreationTimestamp="2025-10-08 14:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:25.190825818 +0000 UTC m=+145.097573310" watchObservedRunningTime="2025-10-08 14:03:25.195563279 +0000 UTC m=+145.102310771" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.198973 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dwjgm" event={"ID":"6fd06952-1b3e-44e6-af19-dfa31965b9ea","Type":"ContainerStarted","Data":"c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.205470 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" podStartSLOduration=124.205443351 podStartE2EDuration="2m4.205443351s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:25.075032999 +0000 UTC m=+144.981780501" watchObservedRunningTime="2025-10-08 14:03:25.205443351 +0000 UTC m=+145.112190843" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.208816 4789 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vzq6g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.209250 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" podUID="5b3f76e2-8cce-47f1-87e5-43be0d07238e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.236013 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" event={"ID":"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6","Type":"ContainerStarted","Data":"814c7addca6a184648176d28a2b8ca4e60a8a8fec04d4585827e87ed98b9cd5a"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.241112 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-dwjgm" podStartSLOduration=124.241094623 podStartE2EDuration="2m4.241094623s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:25.240383303 +0000 UTC m=+145.147130795" watchObservedRunningTime="2025-10-08 14:03:25.241094623 +0000 UTC m=+145.147842115" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.241547 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" podStartSLOduration=124.241543345 podStartE2EDuration="2m4.241543345s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:25.215308723 +0000 UTC m=+145.122056215" watchObservedRunningTime="2025-10-08 14:03:25.241543345 +0000 UTC m=+145.148290837" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.274443 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" event={"ID":"53dc6d10-02c4-48c5-8461-85ed095bcc11","Type":"ContainerStarted","Data":"54ed78a681ff865a3413c2af83fd293a64b13fde1119191de99c1004ae10a457"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.274509 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" event={"ID":"53dc6d10-02c4-48c5-8461-85ed095bcc11","Type":"ContainerStarted","Data":"ad71e3926298540d96a5e9fea6bee92a8daa037862df6466dc8800c3e2a247d0"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.279858 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.281924 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.781895577 +0000 UTC m=+145.688643069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.289230 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" event={"ID":"8e83dccc-9fa2-42d5-ae9a-7941ae00d321","Type":"ContainerStarted","Data":"4a17fe7eb624250107014998ed25da106085c4e93d1e9d92df6154880bc262b1"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.293049 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s" event={"ID":"d79e33d9-6682-4659-8450-bfcca4c46eb6","Type":"ContainerStarted","Data":"bcb52735469616bfd2ea7a027bf530209f4831393c1e9a9823184a362458b5b4"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.295744 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" event={"ID":"5a5f8e9f-467a-4ec3-912a-24dbee088302","Type":"ContainerStarted","Data":"16e4974d1fd8855022518ef1e5726bb4969b2472f01429c909e34c2e9779d9c4"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.311975 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" event={"ID":"e7dc58a5-148e-45bb-9895-a43fe46a0064","Type":"ContainerStarted","Data":"cc9609fcf62114b08e8f129130e8081f24f9f517d8825aaae1c9f6708a8f21e6"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.320668 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p2mjq" event={"ID":"09101a18-6c60-4da9-b517-afb6ae89cb4c","Type":"ContainerStarted","Data":"5fc60dcb06b388278bfe4cb138ef9e1d68eed9f1e20f26a96b4e49694b9ddb61"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.346313 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mvzv5" event={"ID":"6fb689b3-3b0d-476a-ba26-f6836061f366","Type":"ContainerStarted","Data":"1b4ea7b24ffed62131b7fafb8b3e71b2f4fe2bcbd84686dae46d4f40d2ff1be8"} Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.383685 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.386127 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.886105037 +0000 UTC m=+145.792852529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.386126 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mvzv5" podStartSLOduration=124.386106947 podStartE2EDuration="2m4.386106947s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:25.384713198 +0000 UTC m=+145.291460690" watchObservedRunningTime="2025-10-08 14:03:25.386106947 +0000 UTC m=+145.292854439" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.399375 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-lkm5j" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.423386 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-p2mjq" podStartSLOduration=124.423364983 podStartE2EDuration="2m4.423364983s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:25.421252715 +0000 UTC m=+145.328000207" watchObservedRunningTime="2025-10-08 14:03:25.423364983 +0000 UTC m=+145.330112475" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.479642 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" podStartSLOduration=124.479617902 podStartE2EDuration="2m4.479617902s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:25.458640474 +0000 UTC m=+145.365387966" watchObservedRunningTime="2025-10-08 14:03:25.479617902 +0000 UTC m=+145.386365394" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.486256 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.490150 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:25.990116451 +0000 UTC m=+145.896863943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.595016 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.595428 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.095412251 +0000 UTC m=+146.002159743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.652044 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.667364 4789 patch_prober.go:28] interesting pod/router-default-5444994796-p2mjq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 14:03:25 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Oct 08 14:03:25 crc kubenswrapper[4789]: [+]process-running ok Oct 08 14:03:25 crc kubenswrapper[4789]: healthz check failed Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.667418 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p2mjq" podUID="09101a18-6c60-4da9-b517-afb6ae89cb4c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.697815 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.698118 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.198083409 +0000 UTC m=+146.104830911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.698348 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.698719 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.198707806 +0000 UTC m=+146.105455308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.801798 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.802091 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.302043162 +0000 UTC m=+146.208790654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.805853 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.806418 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.306401892 +0000 UTC m=+146.213149384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.906679 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.906955 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.406924011 +0000 UTC m=+146.313671503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:25 crc kubenswrapper[4789]: I1008 14:03:25.907127 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:25 crc kubenswrapper[4789]: E1008 14:03:25.907520 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.407507837 +0000 UTC m=+146.314255329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.008118 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.008302 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.508267002 +0000 UTC m=+146.415014484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.008883 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.009323 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.50929629 +0000 UTC m=+146.416043782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.112851 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.113319 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.613300424 +0000 UTC m=+146.520047916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.159143 4789 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mtgx9 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.159885 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" podUID="119eb4b6-56d0-4cae-803a-56020cda0096" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.223515 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.223771 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.224243 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.724213839 +0000 UTC m=+146.630961331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.328575 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.329432 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.829412827 +0000 UTC m=+146.736160319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.424561 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" event={"ID":"dc66affb-e030-498d-819c-58646b279757","Type":"ContainerStarted","Data":"6b1c6a0cbd5e25c7978c504dfeeec7b1f25411390d70555a444cbafa72025da5"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.432318 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.432391 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" event={"ID":"7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6","Type":"ContainerStarted","Data":"3f1ae406dcfc74980f29a88a0930cfcc99e0c929431a8e4b03028a5a00b1262a"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.432434 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.432678 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.432708 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.433117 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:26.93303584 +0000 UTC m=+146.839783332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.439102 4789 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k4wrp container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.439170 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" podUID="7c592dc6-1a1f-48c5-bee6-e02ccb3f23f6" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.467361 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" event={"ID":"8e83dccc-9fa2-42d5-ae9a-7941ae00d321","Type":"ContainerStarted","Data":"03f8e70ad6e9301e9467f710c5a48294421cd5ed27542e10a872fdbc6ff6bbc9"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.469121 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-prj2m" podStartSLOduration=125.469105084 podStartE2EDuration="2m5.469105084s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.466239265 +0000 UTC m=+146.372986757" watchObservedRunningTime="2025-10-08 14:03:26.469105084 +0000 UTC m=+146.375852576" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.517513 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" podStartSLOduration=125.51749031599999 podStartE2EDuration="2m5.517490316s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.516569511 +0000 UTC m=+146.423317003" watchObservedRunningTime="2025-10-08 14:03:26.517490316 +0000 UTC m=+146.424237798" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.518051 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" event={"ID":"2389a066-e146-4ea8-bb92-d596311b564a","Type":"ContainerStarted","Data":"a111de65cc85e6d65135e4262b26fdc5335bddda34a02b6473abb9cc6425bb89"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.534852 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.542998 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.042949398 +0000 UTC m=+146.949696890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.544639 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.044630654 +0000 UTC m=+146.951378146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.544752 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" event={"ID":"6970b740-c77f-4817-b729-46839dc0698a","Type":"ContainerStarted","Data":"fbe988b00378e4a5b2e9b19539ec12f44e0335276c8293bef2aaac82c2f26535"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.544855 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.568821 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kx5ls" podStartSLOduration=125.568799559 podStartE2EDuration="2m5.568799559s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.566696202 +0000 UTC m=+146.473443694" watchObservedRunningTime="2025-10-08 14:03:26.568799559 +0000 UTC m=+146.475547051" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.589342 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" event={"ID":"6ae8586c-f105-42ff-b845-b4eb6a878694","Type":"ContainerStarted","Data":"456aff7dc1cb350be704edf0265ff832aedf2e5c8d0273368903d1c1544ba388"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.617804 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" event={"ID":"26ee44ad-a6fe-4a2c-bc67-ab856b48a0f6","Type":"ContainerStarted","Data":"8570220c5e2c6348507dea08b21fc5372682c876ee5d34d16f58a446ecf7bf02"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.625277 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dqrvj" podStartSLOduration=125.625261245 podStartE2EDuration="2m5.625261245s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.623025253 +0000 UTC m=+146.529772745" watchObservedRunningTime="2025-10-08 14:03:26.625261245 +0000 UTC m=+146.532008737" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.634849 4789 generic.go:334] "Generic (PLEG): container finished" podID="e7dc58a5-148e-45bb-9895-a43fe46a0064" containerID="78b87375021463ac334c2bb7771ae4db1a92ad875051033f36773c8a9433953d" exitCode=0 Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.634966 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" event={"ID":"e7dc58a5-148e-45bb-9895-a43fe46a0064","Type":"ContainerDied","Data":"78b87375021463ac334c2bb7771ae4db1a92ad875051033f36773c8a9433953d"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.647382 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.648591 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.148568196 +0000 UTC m=+147.055315688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.654751 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-drjkk" podStartSLOduration=125.654733436 podStartE2EDuration="2m5.654733436s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.653671527 +0000 UTC m=+146.560419019" watchObservedRunningTime="2025-10-08 14:03:26.654733436 +0000 UTC m=+146.561480928" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.656699 4789 patch_prober.go:28] interesting pod/router-default-5444994796-p2mjq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 14:03:26 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Oct 08 14:03:26 crc kubenswrapper[4789]: [+]process-running ok Oct 08 14:03:26 crc kubenswrapper[4789]: healthz check failed Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.656779 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p2mjq" podUID="09101a18-6c60-4da9-b517-afb6ae89cb4c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.663718 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" event={"ID":"7191f22d-ffc0-47ff-b226-77f9213db94e","Type":"ContainerStarted","Data":"daa3e6f7f507bb712fc40158b2b4023784c41c3926d6fd704831d6959397ee03"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.664419 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.687306 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" podStartSLOduration=125.687287803 podStartE2EDuration="2m5.687287803s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.683533099 +0000 UTC m=+146.590280591" watchObservedRunningTime="2025-10-08 14:03:26.687287803 +0000 UTC m=+146.594035285" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.691286 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" event={"ID":"17ebe9cd-3f39-42c3-a888-7270e67caa06","Type":"ContainerStarted","Data":"9be4e1e0e2c0e3acebec602f9c47f0d0868e6beee23496cf24783cfa32eca187"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.720795 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" event={"ID":"1c546fd5-9ea3-402e-b837-dca47372c52c","Type":"ContainerStarted","Data":"37598354edcb4a56bf76bf43b32831480b9a5c58ebeb1d4c22e344fe3fc00dc4"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.733655 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" podStartSLOduration=125.733638539 podStartE2EDuration="2m5.733638539s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.717039642 +0000 UTC m=+146.623787134" watchObservedRunningTime="2025-10-08 14:03:26.733638539 +0000 UTC m=+146.640386031" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.747790 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dncbh" event={"ID":"32a50c76-0675-4ff6-b8d2-0239416ed184","Type":"ContainerStarted","Data":"199ef432b5a229fa12f1af5ad1e43a788a631e3e4cc6de8c9929737bf5d27aab"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.749284 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.751819 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.25180554 +0000 UTC m=+147.158553032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.771512 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" event={"ID":"449f550c-38c6-4c62-abc7-2566c2cf1ccc","Type":"ContainerStarted","Data":"41bc9b1d50dbf2c1484677619a438c02ee392497b02fccea130baa7065e18766"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.798460 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" event={"ID":"9d8a129c-9c83-4ed2-a9c5-b017a0bd470b","Type":"ContainerStarted","Data":"b495944958311a4bb0b27637b87d24572c39b330c58b15d5b1c346e1bf31e33c"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.807759 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l7dth" podStartSLOduration=125.80774007 podStartE2EDuration="2m5.80774007s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.804757348 +0000 UTC m=+146.711504840" watchObservedRunningTime="2025-10-08 14:03:26.80774007 +0000 UTC m=+146.714487572" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.809936 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hf8hk" podStartSLOduration=125.8099271 podStartE2EDuration="2m5.8099271s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.773822526 +0000 UTC m=+146.680570028" watchObservedRunningTime="2025-10-08 14:03:26.8099271 +0000 UTC m=+146.716674592" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.835267 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gr58t" event={"ID":"8cf5270f-6d48-40ad-9332-118daa09f473","Type":"ContainerStarted","Data":"cd4acdb4a0163317f14c80a32b0acc1dfd267a6ea554c746923112c5dcf1ff94"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.857898 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.859092 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.359073014 +0000 UTC m=+147.265820506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.879812 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" event={"ID":"5b3f76e2-8cce-47f1-87e5-43be0d07238e","Type":"ContainerStarted","Data":"dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.882269 4789 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vzq6g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.882321 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" podUID="5b3f76e2-8cce-47f1-87e5-43be0d07238e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.914661 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" event={"ID":"73df2e86-6ba0-47d8-bfac-ef6d675808e6","Type":"ContainerStarted","Data":"c866b4a57a8762a162db5fbdc454e12637196b7e0b228bc3fdd050aa87caf9b3"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.914716 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" event={"ID":"73df2e86-6ba0-47d8-bfac-ef6d675808e6","Type":"ContainerStarted","Data":"b3ad40b5722ad7a69a963de513affa0b841549da90161d1a8245a266965a17c1"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.915464 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.929121 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" event={"ID":"1f1b8f79-da9c-4a84-8d59-d8480d3f365e","Type":"ContainerStarted","Data":"14b18826be0338e16a827816150a6ca7b3e68cea1a9fd7776c1086c2081e9803"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.945553 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" event={"ID":"4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9","Type":"ContainerStarted","Data":"5b46923d4fd3bf746254eada68d4a58f1b6dd6990f917ec9bbf7d58022338379"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.946389 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.956070 4789 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-bkblm container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.956119 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" podUID="4ade3df9-f6e4-4ed1-b663-6bb010f6c0a9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.958302 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" event={"ID":"bade0675-1189-4035-9673-2d20edbb729d","Type":"ContainerStarted","Data":"1d7bd21c7ac4ba9399bf14ad33ed542286edba63fc5c3c4c96469e5b55a3b7ef"} Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.959159 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:26 crc kubenswrapper[4789]: E1008 14:03:26.959890 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.45987928 +0000 UTC m=+147.366626772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:26 crc kubenswrapper[4789]: I1008 14:03:26.983175 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" event={"ID":"bc2cc553-6ac7-44a6-948c-c1363836923d","Type":"ContainerStarted","Data":"0be8e11d61f86cf93bb41d9d8a7a0f8d114e1038f66daf9614f8fbccd784f293"} Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.013936 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-dncbh" podStartSLOduration=7.013912989 podStartE2EDuration="7.013912989s" podCreationTimestamp="2025-10-08 14:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.010607577 +0000 UTC m=+146.917355069" watchObservedRunningTime="2025-10-08 14:03:27.013912989 +0000 UTC m=+146.920660481" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.016041 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rbvwr" podStartSLOduration=126.016014556 podStartE2EDuration="2m6.016014556s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:26.86581861 +0000 UTC m=+146.772566102" watchObservedRunningTime="2025-10-08 14:03:27.016014556 +0000 UTC m=+146.922762038" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.026656 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" event={"ID":"53dc6d10-02c4-48c5-8461-85ed095bcc11","Type":"ContainerStarted","Data":"6a038f7d73fdab82afe7f51c7ffd3a79f063d0b7c5c48535599c6943cab7a4a5"} Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.049170 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" event={"ID":"5a5f8e9f-467a-4ec3-912a-24dbee088302","Type":"ContainerStarted","Data":"4ad2bdae2a49d7b43d372ae09c4f128ddb7e3c2b73e81636311a3d863f11089b"} Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.063126 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.064026 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.563975077 +0000 UTC m=+147.470722569 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.071439 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" event={"ID":"60295e4c-18c7-4b3d-acbf-54964ed76081","Type":"ContainerStarted","Data":"a45f6c47e640dfc2a1ca929955d43895974139cba30d87a96a4f848b8484f452"} Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.081471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" event={"ID":"f8317867-1744-48e0-8940-35c3a0c2ab1b","Type":"ContainerStarted","Data":"28899672d40a40ea012e73d2cde8d558adb45ba2406fa5fdf8c9700ca8c41cc6"} Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.105797 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s" event={"ID":"d79e33d9-6682-4659-8450-bfcca4c46eb6","Type":"ContainerStarted","Data":"899610f5d16fa47d5b7a25a3d46e556fa112d78276e79968a4b885c45a173154"} Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.105848 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s" event={"ID":"d79e33d9-6682-4659-8450-bfcca4c46eb6","Type":"ContainerStarted","Data":"25b75a3ec7e36b3a27da4a2a3622cd5f5483d6da0657b570f140a85ba42039c5"} Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.131212 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" event={"ID":"aa675ad1-38c1-49b7-9572-f165efdbd2e9","Type":"ContainerStarted","Data":"667626f1eb0117966bee7e5db3f390e0f5e0798a23592aa9cac3cac8d827334c"} Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.131487 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" event={"ID":"aa675ad1-38c1-49b7-9572-f165efdbd2e9","Type":"ContainerStarted","Data":"485cdbdb1134cdb816f1fec9080d826eea6d454167d04b90ee21b098fe9e4af3"} Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.134641 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mvzv5" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.159181 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvzv5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.159255 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mvzv5" podUID="6fb689b3-3b0d-476a-ba26-f6836061f366" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.159406 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.172899 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.174594 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.674576633 +0000 UTC m=+147.581324205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.274483 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.274677 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.774641518 +0000 UTC m=+147.681389000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.274874 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.277311 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.777295241 +0000 UTC m=+147.684042733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.290347 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" podStartSLOduration=126.29032567 podStartE2EDuration="2m6.29032567s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.241431554 +0000 UTC m=+147.148179046" watchObservedRunningTime="2025-10-08 14:03:27.29032567 +0000 UTC m=+147.197073162" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.346303 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" podStartSLOduration=126.346285501 podStartE2EDuration="2m6.346285501s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.344739439 +0000 UTC m=+147.251486931" watchObservedRunningTime="2025-10-08 14:03:27.346285501 +0000 UTC m=+147.253032993" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.347300 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rmwkt" podStartSLOduration=126.347293739 podStartE2EDuration="2m6.347293739s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.314919908 +0000 UTC m=+147.221667400" watchObservedRunningTime="2025-10-08 14:03:27.347293739 +0000 UTC m=+147.254041231" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.375191 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.375669 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.378739 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.379134 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.879112776 +0000 UTC m=+147.785860268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.391177 4789 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-rh68r container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.391262 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" podUID="e7dc58a5-148e-45bb-9895-a43fe46a0064" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.415462 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-wmqxc" podStartSLOduration=126.415445966 podStartE2EDuration="2m6.415445966s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.409566824 +0000 UTC m=+147.316314316" watchObservedRunningTime="2025-10-08 14:03:27.415445966 +0000 UTC m=+147.322193458" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.417505 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-td8vc" podStartSLOduration=126.417501113 podStartE2EDuration="2m6.417501113s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.370453577 +0000 UTC m=+147.277201079" watchObservedRunningTime="2025-10-08 14:03:27.417501113 +0000 UTC m=+147.324248595" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.459329 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xpczn" podStartSLOduration=126.459315555 podStartE2EDuration="2m6.459315555s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.458025509 +0000 UTC m=+147.364773001" watchObservedRunningTime="2025-10-08 14:03:27.459315555 +0000 UTC m=+147.366063047" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.482137 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.482652 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:27.982635427 +0000 UTC m=+147.889382919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.571894 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-g9tdd" podStartSLOduration=126.571871004 podStartE2EDuration="2m6.571871004s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.499691946 +0000 UTC m=+147.406439438" watchObservedRunningTime="2025-10-08 14:03:27.571871004 +0000 UTC m=+147.478618496" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.584048 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.584455 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.08443333 +0000 UTC m=+147.991180822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.654940 4789 patch_prober.go:28] interesting pod/router-default-5444994796-p2mjq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 14:03:27 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Oct 08 14:03:27 crc kubenswrapper[4789]: [+]process-running ok Oct 08 14:03:27 crc kubenswrapper[4789]: healthz check failed Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.655019 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p2mjq" podUID="09101a18-6c60-4da9-b517-afb6ae89cb4c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.661199 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7pwrl" podStartSLOduration=126.661181034 podStartE2EDuration="2m6.661181034s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.626219421 +0000 UTC m=+147.532966913" watchObservedRunningTime="2025-10-08 14:03:27.661181034 +0000 UTC m=+147.567928516" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.686364 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.686717 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.186705077 +0000 UTC m=+148.093452569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.699214 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qlf4j" podStartSLOduration=126.699196501 podStartE2EDuration="2m6.699196501s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.663896449 +0000 UTC m=+147.570643941" watchObservedRunningTime="2025-10-08 14:03:27.699196501 +0000 UTC m=+147.605943983" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.699745 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ljh2s" podStartSLOduration=126.699740286 podStartE2EDuration="2m6.699740286s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.69443506 +0000 UTC m=+147.601182542" watchObservedRunningTime="2025-10-08 14:03:27.699740286 +0000 UTC m=+147.606487778" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.750824 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" podStartSLOduration=126.750807393 podStartE2EDuration="2m6.750807393s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.750454043 +0000 UTC m=+147.657201535" watchObservedRunningTime="2025-10-08 14:03:27.750807393 +0000 UTC m=+147.657554885" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.790613 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.790944 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.290924277 +0000 UTC m=+148.197671769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.791643 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vtftw" podStartSLOduration=128.791623697 podStartE2EDuration="2m8.791623697s" podCreationTimestamp="2025-10-08 14:01:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:27.791229006 +0000 UTC m=+147.697976488" watchObservedRunningTime="2025-10-08 14:03:27.791623697 +0000 UTC m=+147.698371189" Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.892774 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.893146 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.393129952 +0000 UTC m=+148.299877444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.994140 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.994280 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.494255287 +0000 UTC m=+148.401002779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:27 crc kubenswrapper[4789]: I1008 14:03:27.994382 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:27 crc kubenswrapper[4789]: E1008 14:03:27.994700 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.49469284 +0000 UTC m=+148.401440322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.095074 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.095259 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.595227608 +0000 UTC m=+148.501975100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.095405 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.095778 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.595770243 +0000 UTC m=+148.502517735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.134499 4789 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-84p6z container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.134557 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" podUID="172550ec-f5a7-4fb2-8c89-8ec82e2775f7" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.137828 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" event={"ID":"6970b740-c77f-4817-b729-46839dc0698a","Type":"ContainerStarted","Data":"de2913492f72a8aaafba365732a6f4a4f30b09eff05c4077f5887cb97f6649d6"} Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.139390 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-xs8k5" event={"ID":"5a5f8e9f-467a-4ec3-912a-24dbee088302","Type":"ContainerStarted","Data":"67b6599a317955371afee77dd6f13444c089244af34201db7a19bb25d651b15a"} Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.141206 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" event={"ID":"e7dc58a5-148e-45bb-9895-a43fe46a0064","Type":"ContainerStarted","Data":"e5d8c4c92cdfa2016077da441dbd62014d2a29965177828c15ffbc6b53aba456"} Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.142845 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gr58t" event={"ID":"8cf5270f-6d48-40ad-9332-118daa09f473","Type":"ContainerStarted","Data":"d9e927e6d1c5c51e959ce4636d629f9528949b9627a76e8d4a3605e586b64c1d"} Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.143234 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.144593 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96vg6" event={"ID":"2b2efb14-dde5-4a97-9661-9ff6ec231414","Type":"ContainerStarted","Data":"9aaab96d7dda0ea0be2ec7b037152aa950623b08964d7c35d98601fdf59ef495"} Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.146816 4789 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vzq6g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.146851 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" podUID="5b3f76e2-8cce-47f1-87e5-43be0d07238e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.147567 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvzv5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.147594 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mvzv5" podUID="6fb689b3-3b0d-476a-ba26-f6836061f366" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.155898 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k4wrp" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.176173 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-bkblm" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.195972 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.196405 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.696374564 +0000 UTC m=+148.603122056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.275602 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" podStartSLOduration=128.275581326 podStartE2EDuration="2m8.275581326s" podCreationTimestamp="2025-10-08 14:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:28.232682634 +0000 UTC m=+148.139430126" watchObservedRunningTime="2025-10-08 14:03:28.275581326 +0000 UTC m=+148.182328818" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.299098 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.306772 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.806753404 +0000 UTC m=+148.713500896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.319041 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-gr58t" podStartSLOduration=9.31896198 podStartE2EDuration="9.31896198s" podCreationTimestamp="2025-10-08 14:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:28.317566302 +0000 UTC m=+148.224313794" watchObservedRunningTime="2025-10-08 14:03:28.31896198 +0000 UTC m=+148.225709472" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.406745 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.407091 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:28.907069937 +0000 UTC m=+148.813817429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.508161 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.508605 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.008586823 +0000 UTC m=+148.915334315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.609836 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.610073 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.110039007 +0000 UTC m=+149.016786499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.654179 4789 patch_prober.go:28] interesting pod/router-default-5444994796-p2mjq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 14:03:28 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Oct 08 14:03:28 crc kubenswrapper[4789]: [+]process-running ok Oct 08 14:03:28 crc kubenswrapper[4789]: healthz check failed Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.654237 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p2mjq" podUID="09101a18-6c60-4da9-b517-afb6ae89cb4c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.711929 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.712415 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.212400106 +0000 UTC m=+149.119147598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.812891 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.813161 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.3131276 +0000 UTC m=+149.219875092 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.813520 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.813860 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.31384468 +0000 UTC m=+149.220592172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.851038 4789 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.914783 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.915212 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.415188171 +0000 UTC m=+149.321935663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.915287 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:28 crc kubenswrapper[4789]: E1008 14:03:28.915699 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 14:03:29.415679475 +0000 UTC m=+149.322426967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mc249" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.933793 4789 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-08T14:03:28.851065815Z","Handler":null,"Name":""} Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.967184 4789 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.967235 4789 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.994852 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-72nk5"] Oct 08 14:03:28 crc kubenswrapper[4789]: I1008 14:03:28.996858 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.005317 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zmdbz" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.005833 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.010679 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72nk5"] Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.016626 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.049961 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.118537 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-catalog-content\") pod \"community-operators-72nk5\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.118621 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.118645 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-utilities\") pod \"community-operators-72nk5\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.118683 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dwfh\" (UniqueName: \"kubernetes.io/projected/a5e65840-1d44-4cb3-bd42-3bc0048f7274-kube-api-access-5dwfh\") pod \"community-operators-72nk5\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.147525 4789 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.147571 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.151063 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96vg6" event={"ID":"2b2efb14-dde5-4a97-9661-9ff6ec231414","Type":"ContainerStarted","Data":"bf9d58bd9f43511da13590f970494d6aae33dd9f7d01e5a0a94d029d015df831"} Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.151119 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96vg6" event={"ID":"2b2efb14-dde5-4a97-9661-9ff6ec231414","Type":"ContainerStarted","Data":"5c21e816da45222398359b282c75cba034cfab106ca0f90a4d9034135a5b3e27"} Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.153202 4789 generic.go:334] "Generic (PLEG): container finished" podID="2389a066-e146-4ea8-bb92-d596311b564a" containerID="a111de65cc85e6d65135e4262b26fdc5335bddda34a02b6473abb9cc6425bb89" exitCode=0 Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.154338 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" event={"ID":"2389a066-e146-4ea8-bb92-d596311b564a","Type":"ContainerDied","Data":"a111de65cc85e6d65135e4262b26fdc5335bddda34a02b6473abb9cc6425bb89"} Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.208312 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mc249\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.221927 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-utilities\") pod \"community-operators-72nk5\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.222398 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-utilities\") pod \"community-operators-72nk5\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.222890 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dwfh\" (UniqueName: \"kubernetes.io/projected/a5e65840-1d44-4cb3-bd42-3bc0048f7274-kube-api-access-5dwfh\") pod \"community-operators-72nk5\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.224307 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-catalog-content\") pod \"community-operators-72nk5\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.224583 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-catalog-content\") pod \"community-operators-72nk5\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.241205 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.251828 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dwfh\" (UniqueName: \"kubernetes.io/projected/a5e65840-1d44-4cb3-bd42-3bc0048f7274-kube-api-access-5dwfh\") pod \"community-operators-72nk5\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.322879 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.397837 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rqnlm"] Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.399315 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.411849 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqnlm"] Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.426479 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2676f\" (UniqueName: \"kubernetes.io/projected/5e523982-53d8-4082-a367-d99a110d544f-kube-api-access-2676f\") pod \"community-operators-rqnlm\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.426563 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-utilities\") pod \"community-operators-rqnlm\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.426583 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-catalog-content\") pod \"community-operators-rqnlm\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.528487 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2676f\" (UniqueName: \"kubernetes.io/projected/5e523982-53d8-4082-a367-d99a110d544f-kube-api-access-2676f\") pod \"community-operators-rqnlm\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.528612 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-utilities\") pod \"community-operators-rqnlm\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.528639 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-catalog-content\") pod \"community-operators-rqnlm\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.529923 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-utilities\") pod \"community-operators-rqnlm\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.530002 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-catalog-content\") pod \"community-operators-rqnlm\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.557905 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2676f\" (UniqueName: \"kubernetes.io/projected/5e523982-53d8-4082-a367-d99a110d544f-kube-api-access-2676f\") pod \"community-operators-rqnlm\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.588927 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mc249"] Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.598869 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rswj8"] Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.599757 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.606120 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.628912 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-utilities\") pod \"certified-operators-rswj8\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.628966 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-catalog-content\") pod \"certified-operators-rswj8\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.629022 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rbws\" (UniqueName: \"kubernetes.io/projected/2908d4c0-bbe0-436b-b806-ca29cf20afcf-kube-api-access-2rbws\") pod \"certified-operators-rswj8\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.633114 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rswj8"] Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.662982 4789 patch_prober.go:28] interesting pod/router-default-5444994796-p2mjq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 14:03:29 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Oct 08 14:03:29 crc kubenswrapper[4789]: [+]process-running ok Oct 08 14:03:29 crc kubenswrapper[4789]: healthz check failed Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.663076 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p2mjq" podUID="09101a18-6c60-4da9-b517-afb6ae89cb4c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.696706 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72nk5"] Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.729646 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-utilities\") pod \"certified-operators-rswj8\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.729708 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.729733 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-catalog-content\") pod \"certified-operators-rswj8\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.729759 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.729779 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.729802 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rbws\" (UniqueName: \"kubernetes.io/projected/2908d4c0-bbe0-436b-b806-ca29cf20afcf-kube-api-access-2rbws\") pod \"certified-operators-rswj8\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.729853 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.734819 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.735554 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-catalog-content\") pod \"certified-operators-rswj8\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.735789 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-utilities\") pod \"certified-operators-rswj8\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.737360 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.739749 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.744646 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.744665 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.763315 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.769955 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rbws\" (UniqueName: \"kubernetes.io/projected/2908d4c0-bbe0-436b-b806-ca29cf20afcf-kube-api-access-2rbws\") pod \"certified-operators-rswj8\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.770419 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.789485 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p9cvl"] Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.791395 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.802298 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9cvl"] Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.932982 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69nvc\" (UniqueName: \"kubernetes.io/projected/dc0408a7-643a-4260-9d54-382cdb73caca-kube-api-access-69nvc\") pod \"certified-operators-p9cvl\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.933038 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-catalog-content\") pod \"certified-operators-p9cvl\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.933058 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-utilities\") pod \"certified-operators-p9cvl\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.953630 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 14:03:29 crc kubenswrapper[4789]: I1008 14:03:29.976630 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.019176 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqnlm"] Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.035241 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69nvc\" (UniqueName: \"kubernetes.io/projected/dc0408a7-643a-4260-9d54-382cdb73caca-kube-api-access-69nvc\") pod \"certified-operators-p9cvl\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.035285 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-catalog-content\") pod \"certified-operators-p9cvl\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.035303 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-utilities\") pod \"certified-operators-p9cvl\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.035797 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-utilities\") pod \"certified-operators-p9cvl\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.036124 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-catalog-content\") pod \"certified-operators-p9cvl\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.063440 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69nvc\" (UniqueName: \"kubernetes.io/projected/dc0408a7-643a-4260-9d54-382cdb73caca-kube-api-access-69nvc\") pod \"certified-operators-p9cvl\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.115671 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.184698 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-96vg6" event={"ID":"2b2efb14-dde5-4a97-9661-9ff6ec231414","Type":"ContainerStarted","Data":"d49cd778b517c6d290b6ea8f262a090f7fc18e150c2b51e8449c5a2f41b982da"} Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.187636 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8329dfa6c35fd67b1dd237fa53ac4201ad6955814998b2ee5cbf953587aea6f2"} Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.189551 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" event={"ID":"d124aeeb-8b80-4c21-b9c5-617be3ed17f6","Type":"ContainerStarted","Data":"b5f4ced7f877ef91a2daca300c3de8f3d2e9a1495e1800c87524af7796a9be59"} Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.189575 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" event={"ID":"d124aeeb-8b80-4c21-b9c5-617be3ed17f6","Type":"ContainerStarted","Data":"f77aa9982285ad5a0dda5c6b4d7aee213deb7f4e00eda7baa3e35360245e5de5"} Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.190443 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.192602 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72nk5" event={"ID":"a5e65840-1d44-4cb3-bd42-3bc0048f7274","Type":"ContainerStarted","Data":"006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6"} Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.192644 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72nk5" event={"ID":"a5e65840-1d44-4cb3-bd42-3bc0048f7274","Type":"ContainerStarted","Data":"00806708096f1a713137a29d99b0c1ab338848db74806ad452e2d9554ea2c18a"} Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.194140 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"14e8d413385064ee4c2c3c460ba0b0e08cbff93f5852c541ca21b766d63c2898"} Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.196430 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnlm" event={"ID":"5e523982-53d8-4082-a367-d99a110d544f","Type":"ContainerStarted","Data":"e34663c74d429a97bfeacee83c4651a3d6bd72d8ef43a024af49351ffb2e356c"} Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.199424 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rswj8"] Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.206921 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-96vg6" podStartSLOduration=11.206902287 podStartE2EDuration="11.206902287s" podCreationTimestamp="2025-10-08 14:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:30.205917049 +0000 UTC m=+150.112664541" watchObservedRunningTime="2025-10-08 14:03:30.206902287 +0000 UTC m=+150.113649779" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.229338 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" podStartSLOduration=129.229316204 podStartE2EDuration="2m9.229316204s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:30.226962169 +0000 UTC m=+150.133709661" watchObservedRunningTime="2025-10-08 14:03:30.229316204 +0000 UTC m=+150.136063686" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.474528 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9cvl"] Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.566598 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.647849 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2389a066-e146-4ea8-bb92-d596311b564a-secret-volume\") pod \"2389a066-e146-4ea8-bb92-d596311b564a\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.647920 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2389a066-e146-4ea8-bb92-d596311b564a-config-volume\") pod \"2389a066-e146-4ea8-bb92-d596311b564a\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.647957 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r67d6\" (UniqueName: \"kubernetes.io/projected/2389a066-e146-4ea8-bb92-d596311b564a-kube-api-access-r67d6\") pod \"2389a066-e146-4ea8-bb92-d596311b564a\" (UID: \"2389a066-e146-4ea8-bb92-d596311b564a\") " Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.648910 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2389a066-e146-4ea8-bb92-d596311b564a-config-volume" (OuterVolumeSpecName: "config-volume") pod "2389a066-e146-4ea8-bb92-d596311b564a" (UID: "2389a066-e146-4ea8-bb92-d596311b564a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.654169 4789 patch_prober.go:28] interesting pod/router-default-5444994796-p2mjq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 14:03:30 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Oct 08 14:03:30 crc kubenswrapper[4789]: [+]process-running ok Oct 08 14:03:30 crc kubenswrapper[4789]: healthz check failed Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.654239 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p2mjq" podUID="09101a18-6c60-4da9-b517-afb6ae89cb4c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.655429 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2389a066-e146-4ea8-bb92-d596311b564a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2389a066-e146-4ea8-bb92-d596311b564a" (UID: "2389a066-e146-4ea8-bb92-d596311b564a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.664746 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2389a066-e146-4ea8-bb92-d596311b564a-kube-api-access-r67d6" (OuterVolumeSpecName: "kube-api-access-r67d6") pod "2389a066-e146-4ea8-bb92-d596311b564a" (UID: "2389a066-e146-4ea8-bb92-d596311b564a"). InnerVolumeSpecName "kube-api-access-r67d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.737719 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.749672 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2389a066-e146-4ea8-bb92-d596311b564a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.749707 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2389a066-e146-4ea8-bb92-d596311b564a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.749720 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r67d6\" (UniqueName: \"kubernetes.io/projected/2389a066-e146-4ea8-bb92-d596311b564a-kube-api-access-r67d6\") on node \"crc\" DevicePath \"\"" Oct 08 14:03:30 crc kubenswrapper[4789]: I1008 14:03:30.754423 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.201567 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c303f173e6cca5cd2c55c6ae3a05aacc6d7ad64bdb904ab9ea0baab46e81c870"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.201878 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"db554abfe92d635548214aa6bb0cdbb6e05572602b23cdf08ae01b673d26b867"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.203215 4789 generic.go:334] "Generic (PLEG): container finished" podID="dc0408a7-643a-4260-9d54-382cdb73caca" containerID="7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb" exitCode=0 Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.203279 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9cvl" event={"ID":"dc0408a7-643a-4260-9d54-382cdb73caca","Type":"ContainerDied","Data":"7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.203347 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9cvl" event={"ID":"dc0408a7-643a-4260-9d54-382cdb73caca","Type":"ContainerStarted","Data":"2f9fa3ab463bb90792a3c8f602dc7280982fde3c984fc4b7f025758759a1fa8e"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.205395 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.206452 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4054aba69ff143afd82e0e157264e5110dc99a16046633947a30c7a5b165cf77"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.206798 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.209423 4789 generic.go:334] "Generic (PLEG): container finished" podID="5e523982-53d8-4082-a367-d99a110d544f" containerID="0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce" exitCode=0 Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.210013 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnlm" event={"ID":"5e523982-53d8-4082-a367-d99a110d544f","Type":"ContainerDied","Data":"0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.211550 4789 generic.go:334] "Generic (PLEG): container finished" podID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerID="00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185" exitCode=0 Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.211663 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rswj8" event={"ID":"2908d4c0-bbe0-436b-b806-ca29cf20afcf","Type":"ContainerDied","Data":"00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.211722 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rswj8" event={"ID":"2908d4c0-bbe0-436b-b806-ca29cf20afcf","Type":"ContainerStarted","Data":"cad15f30d5ee880ae26d8d1aca426e0b81871e3ea063f1d6ce3f4e9ef7a6b176"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.215077 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" event={"ID":"2389a066-e146-4ea8-bb92-d596311b564a","Type":"ContainerDied","Data":"3b44d2293810e705c7b1f5e97e9bced29d6a0df57d91af766ea28f9a09936bf6"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.215163 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b44d2293810e705c7b1f5e97e9bced29d6a0df57d91af766ea28f9a09936bf6" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.215127 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.220769 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"672ff08e2526d6bf09e93a74a883ea28bcad6fd1397d3c6cfa2f508ccada3561"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.222937 4789 generic.go:334] "Generic (PLEG): container finished" podID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerID="006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6" exitCode=0 Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.223138 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72nk5" event={"ID":"a5e65840-1d44-4cb3-bd42-3bc0048f7274","Type":"ContainerDied","Data":"006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6"} Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.390482 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z84nj"] Oct 08 14:03:31 crc kubenswrapper[4789]: E1008 14:03:31.390671 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2389a066-e146-4ea8-bb92-d596311b564a" containerName="collect-profiles" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.390684 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2389a066-e146-4ea8-bb92-d596311b564a" containerName="collect-profiles" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.390786 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2389a066-e146-4ea8-bb92-d596311b564a" containerName="collect-profiles" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.391513 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.398667 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.402646 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z84nj"] Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.566403 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-catalog-content\") pod \"redhat-marketplace-z84nj\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.566452 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-utilities\") pod \"redhat-marketplace-z84nj\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.566484 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d59rc\" (UniqueName: \"kubernetes.io/projected/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-kube-api-access-d59rc\") pod \"redhat-marketplace-z84nj\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.655014 4789 patch_prober.go:28] interesting pod/router-default-5444994796-p2mjq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 14:03:31 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Oct 08 14:03:31 crc kubenswrapper[4789]: [+]process-running ok Oct 08 14:03:31 crc kubenswrapper[4789]: healthz check failed Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.655085 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p2mjq" podUID="09101a18-6c60-4da9-b517-afb6ae89cb4c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.667406 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-utilities\") pod \"redhat-marketplace-z84nj\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.667519 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d59rc\" (UniqueName: \"kubernetes.io/projected/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-kube-api-access-d59rc\") pod \"redhat-marketplace-z84nj\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.667799 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-catalog-content\") pod \"redhat-marketplace-z84nj\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.667896 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-utilities\") pod \"redhat-marketplace-z84nj\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.668571 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-catalog-content\") pod \"redhat-marketplace-z84nj\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.695317 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d59rc\" (UniqueName: \"kubernetes.io/projected/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-kube-api-access-d59rc\") pod \"redhat-marketplace-z84nj\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.731461 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.791605 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcr8"] Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.793093 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.812031 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcr8"] Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.951350 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z84nj"] Oct 08 14:03:31 crc kubenswrapper[4789]: W1008 14:03:31.955737 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35d2ebf2_6bd4_4adf_872d_d241a2519b0d.slice/crio-b47b8039f0281bb7fe525583194bb9e3adb28b4abc2d189078f4e1cb3d7c570c WatchSource:0}: Error finding container b47b8039f0281bb7fe525583194bb9e3adb28b4abc2d189078f4e1cb3d7c570c: Status 404 returned error can't find the container with id b47b8039f0281bb7fe525583194bb9e3adb28b4abc2d189078f4e1cb3d7c570c Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.972810 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cw5x\" (UniqueName: \"kubernetes.io/projected/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-kube-api-access-7cw5x\") pod \"redhat-marketplace-lzcr8\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.972913 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-utilities\") pod \"redhat-marketplace-lzcr8\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:31 crc kubenswrapper[4789]: I1008 14:03:31.972943 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-catalog-content\") pod \"redhat-marketplace-lzcr8\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:31.993625 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:31.994378 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:31.997304 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:31.997896 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.001003 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.073929 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-utilities\") pod \"redhat-marketplace-lzcr8\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.074556 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-catalog-content\") pod \"redhat-marketplace-lzcr8\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.074669 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-utilities\") pod \"redhat-marketplace-lzcr8\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.074680 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cw5x\" (UniqueName: \"kubernetes.io/projected/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-kube-api-access-7cw5x\") pod \"redhat-marketplace-lzcr8\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.075060 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-catalog-content\") pod \"redhat-marketplace-lzcr8\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.094807 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cw5x\" (UniqueName: \"kubernetes.io/projected/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-kube-api-access-7cw5x\") pod \"redhat-marketplace-lzcr8\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.115483 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.135053 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.135906 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.150808 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.152449 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.160308 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.180624 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12692816-f57d-4308-a946-261fee1ee271-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"12692816-f57d-4308-a946-261fee1ee271\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.181289 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12692816-f57d-4308-a946-261fee1ee271-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"12692816-f57d-4308-a946-261fee1ee271\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.193589 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvzv5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.193652 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mvzv5" podUID="6fb689b3-3b0d-476a-ba26-f6836061f366" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.193688 4789 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvzv5 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.193763 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mvzv5" podUID="6fb689b3-3b0d-476a-ba26-f6836061f366" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.202555 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.203602 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.208360 4789 patch_prober.go:28] interesting pod/console-f9d7485db-dwjgm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.208598 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-dwjgm" podUID="6fd06952-1b3e-44e6-af19-dfa31965b9ea" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.230455 4789 generic.go:334] "Generic (PLEG): container finished" podID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerID="d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288" exitCode=0 Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.230865 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z84nj" event={"ID":"35d2ebf2-6bd4-4adf-872d-d241a2519b0d","Type":"ContainerDied","Data":"d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288"} Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.232787 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z84nj" event={"ID":"35d2ebf2-6bd4-4adf-872d-d241a2519b0d","Type":"ContainerStarted","Data":"b47b8039f0281bb7fe525583194bb9e3adb28b4abc2d189078f4e1cb3d7c570c"} Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.239428 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kch4x" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.283396 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12692816-f57d-4308-a946-261fee1ee271-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"12692816-f57d-4308-a946-261fee1ee271\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.283437 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12692816-f57d-4308-a946-261fee1ee271-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"12692816-f57d-4308-a946-261fee1ee271\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.284189 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12692816-f57d-4308-a946-261fee1ee271-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"12692816-f57d-4308-a946-261fee1ee271\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.333863 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12692816-f57d-4308-a946-261fee1ee271-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"12692816-f57d-4308-a946-261fee1ee271\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.341802 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.397343 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vn2fk"] Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.399486 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.400393 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.406150 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.410403 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-84p6z" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.410795 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rh68r" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.448524 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vn2fk"] Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.481807 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcr8"] Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.498945 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-utilities\") pod \"redhat-operators-vn2fk\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.499078 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-catalog-content\") pod \"redhat-operators-vn2fk\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.499110 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjbkg\" (UniqueName: \"kubernetes.io/projected/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-kube-api-access-gjbkg\") pod \"redhat-operators-vn2fk\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.599864 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-utilities\") pod \"redhat-operators-vn2fk\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.600282 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-catalog-content\") pod \"redhat-operators-vn2fk\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.600307 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjbkg\" (UniqueName: \"kubernetes.io/projected/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-kube-api-access-gjbkg\") pod \"redhat-operators-vn2fk\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.601160 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-utilities\") pod \"redhat-operators-vn2fk\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.601388 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-catalog-content\") pod \"redhat-operators-vn2fk\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.645230 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjbkg\" (UniqueName: \"kubernetes.io/projected/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-kube-api-access-gjbkg\") pod \"redhat-operators-vn2fk\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.650768 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.664233 4789 patch_prober.go:28] interesting pod/router-default-5444994796-p2mjq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 14:03:32 crc kubenswrapper[4789]: [-]has-synced failed: reason withheld Oct 08 14:03:32 crc kubenswrapper[4789]: [+]process-running ok Oct 08 14:03:32 crc kubenswrapper[4789]: healthz check failed Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.664294 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p2mjq" podUID="09101a18-6c60-4da9-b517-afb6ae89cb4c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.671886 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.803820 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9z9p6"] Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.805730 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.821757 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9z9p6"] Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.841893 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.912878 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb6sq\" (UniqueName: \"kubernetes.io/projected/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-kube-api-access-bb6sq\") pod \"redhat-operators-9z9p6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.913396 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-catalog-content\") pod \"redhat-operators-9z9p6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:32 crc kubenswrapper[4789]: I1008 14:03:32.913471 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-utilities\") pod \"redhat-operators-9z9p6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.016720 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-utilities\") pod \"redhat-operators-9z9p6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.016820 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb6sq\" (UniqueName: \"kubernetes.io/projected/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-kube-api-access-bb6sq\") pod \"redhat-operators-9z9p6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.016887 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-catalog-content\") pod \"redhat-operators-9z9p6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.017611 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-utilities\") pod \"redhat-operators-9z9p6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.017829 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-catalog-content\") pod \"redhat-operators-9z9p6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.040717 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb6sq\" (UniqueName: \"kubernetes.io/projected/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-kube-api-access-bb6sq\") pod \"redhat-operators-9z9p6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.143825 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.153395 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:03:33 crc kubenswrapper[4789]: W1008 14:03:33.172858 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod12692816_f57d_4308_a946_261fee1ee271.slice/crio-036a144be9a8c8aac0fbc95424d40aee9b9ae7feb8a0d57e4b79e79db4a32ad5 WatchSource:0}: Error finding container 036a144be9a8c8aac0fbc95424d40aee9b9ae7feb8a0d57e4b79e79db4a32ad5: Status 404 returned error can't find the container with id 036a144be9a8c8aac0fbc95424d40aee9b9ae7feb8a0d57e4b79e79db4a32ad5 Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.269764 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vn2fk"] Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.300841 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"12692816-f57d-4308-a946-261fee1ee271","Type":"ContainerStarted","Data":"036a144be9a8c8aac0fbc95424d40aee9b9ae7feb8a0d57e4b79e79db4a32ad5"} Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.324976 4789 generic.go:334] "Generic (PLEG): container finished" podID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerID="f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b" exitCode=0 Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.325892 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcr8" event={"ID":"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a","Type":"ContainerDied","Data":"f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b"} Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.325926 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcr8" event={"ID":"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a","Type":"ContainerStarted","Data":"2058556dfcd0f84ddccc8815917bfc4d03ed8caed386fa6f59feecf4b6fb35c4"} Oct 08 14:03:33 crc kubenswrapper[4789]: W1008 14:03:33.341788 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68b331a3_a0a4_4fc8_bcd0_7d180f2a86be.slice/crio-a8a62359b1b7d54218ec3d87223ebd5622772ea47a4a2927ba711b10c158c2a1 WatchSource:0}: Error finding container a8a62359b1b7d54218ec3d87223ebd5622772ea47a4a2927ba711b10c158c2a1: Status 404 returned error can't find the container with id a8a62359b1b7d54218ec3d87223ebd5622772ea47a4a2927ba711b10c158c2a1 Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.660338 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.665703 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-p2mjq" Oct 08 14:03:33 crc kubenswrapper[4789]: I1008 14:03:33.992320 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9z9p6"] Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.304916 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.307544 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.309861 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.310187 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.320475 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.377712 4789 generic.go:334] "Generic (PLEG): container finished" podID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerID="814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d" exitCode=0 Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.377784 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2fk" event={"ID":"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be","Type":"ContainerDied","Data":"814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d"} Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.377820 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2fk" event={"ID":"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be","Type":"ContainerStarted","Data":"a8a62359b1b7d54218ec3d87223ebd5622772ea47a4a2927ba711b10c158c2a1"} Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.387743 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z9p6" event={"ID":"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6","Type":"ContainerStarted","Data":"67ba6eb3b1cd3dc9e069f54a3ddbc91bda15455a07b6ee0ed59b896cab1508f2"} Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.387770 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z9p6" event={"ID":"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6","Type":"ContainerStarted","Data":"70a3c79d6120b673d9c20bfceddbd038e17aa087df781d5a555c43f71e7c8e95"} Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.396233 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"12692816-f57d-4308-a946-261fee1ee271","Type":"ContainerStarted","Data":"1f3020651d0bf309c6a7f7f17e3f6af3df095115176cccec4479971ed56f84fd"} Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.444164 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.444328 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.545464 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.545529 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.546916 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.569531 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:34 crc kubenswrapper[4789]: I1008 14:03:34.642679 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:35 crc kubenswrapper[4789]: I1008 14:03:35.324584 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.324529211 podStartE2EDuration="4.324529211s" podCreationTimestamp="2025-10-08 14:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:34.443487916 +0000 UTC m=+154.350235408" watchObservedRunningTime="2025-10-08 14:03:35.324529211 +0000 UTC m=+155.231276713" Oct 08 14:03:35 crc kubenswrapper[4789]: I1008 14:03:35.325476 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 14:03:35 crc kubenswrapper[4789]: W1008 14:03:35.358709 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9d8f6fac_c3df_4016_b978_ff0e17ea642f.slice/crio-338ce4bbdfc997917e3ce4dfe6b4fa41c4ed594be8536e01a7cb69fd36c2c817 WatchSource:0}: Error finding container 338ce4bbdfc997917e3ce4dfe6b4fa41c4ed594be8536e01a7cb69fd36c2c817: Status 404 returned error can't find the container with id 338ce4bbdfc997917e3ce4dfe6b4fa41c4ed594be8536e01a7cb69fd36c2c817 Oct 08 14:03:35 crc kubenswrapper[4789]: I1008 14:03:35.406284 4789 generic.go:334] "Generic (PLEG): container finished" podID="12692816-f57d-4308-a946-261fee1ee271" containerID="1f3020651d0bf309c6a7f7f17e3f6af3df095115176cccec4479971ed56f84fd" exitCode=0 Oct 08 14:03:35 crc kubenswrapper[4789]: I1008 14:03:35.406358 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"12692816-f57d-4308-a946-261fee1ee271","Type":"ContainerDied","Data":"1f3020651d0bf309c6a7f7f17e3f6af3df095115176cccec4479971ed56f84fd"} Oct 08 14:03:35 crc kubenswrapper[4789]: I1008 14:03:35.411327 4789 generic.go:334] "Generic (PLEG): container finished" podID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerID="67ba6eb3b1cd3dc9e069f54a3ddbc91bda15455a07b6ee0ed59b896cab1508f2" exitCode=0 Oct 08 14:03:35 crc kubenswrapper[4789]: I1008 14:03:35.411438 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z9p6" event={"ID":"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6","Type":"ContainerDied","Data":"67ba6eb3b1cd3dc9e069f54a3ddbc91bda15455a07b6ee0ed59b896cab1508f2"} Oct 08 14:03:35 crc kubenswrapper[4789]: I1008 14:03:35.419272 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9d8f6fac-c3df-4016-b978-ff0e17ea642f","Type":"ContainerStarted","Data":"338ce4bbdfc997917e3ce4dfe6b4fa41c4ed594be8536e01a7cb69fd36c2c817"} Oct 08 14:03:36 crc kubenswrapper[4789]: I1008 14:03:36.733302 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:36 crc kubenswrapper[4789]: I1008 14:03:36.776031 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12692816-f57d-4308-a946-261fee1ee271-kubelet-dir\") pod \"12692816-f57d-4308-a946-261fee1ee271\" (UID: \"12692816-f57d-4308-a946-261fee1ee271\") " Oct 08 14:03:36 crc kubenswrapper[4789]: I1008 14:03:36.776197 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12692816-f57d-4308-a946-261fee1ee271-kube-api-access\") pod \"12692816-f57d-4308-a946-261fee1ee271\" (UID: \"12692816-f57d-4308-a946-261fee1ee271\") " Oct 08 14:03:36 crc kubenswrapper[4789]: I1008 14:03:36.776182 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12692816-f57d-4308-a946-261fee1ee271-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "12692816-f57d-4308-a946-261fee1ee271" (UID: "12692816-f57d-4308-a946-261fee1ee271"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:03:36 crc kubenswrapper[4789]: I1008 14:03:36.776723 4789 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12692816-f57d-4308-a946-261fee1ee271-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 14:03:36 crc kubenswrapper[4789]: I1008 14:03:36.800423 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12692816-f57d-4308-a946-261fee1ee271-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "12692816-f57d-4308-a946-261fee1ee271" (UID: "12692816-f57d-4308-a946-261fee1ee271"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:03:36 crc kubenswrapper[4789]: I1008 14:03:36.882886 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12692816-f57d-4308-a946-261fee1ee271-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 14:03:37 crc kubenswrapper[4789]: I1008 14:03:37.473068 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9d8f6fac-c3df-4016-b978-ff0e17ea642f","Type":"ContainerStarted","Data":"d9e101585d9b4534e3297903c9578627ab41e00d5ef68dcabb8066288fece343"} Oct 08 14:03:37 crc kubenswrapper[4789]: I1008 14:03:37.492883 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.492850359 podStartE2EDuration="3.492850359s" podCreationTimestamp="2025-10-08 14:03:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:03:37.486211796 +0000 UTC m=+157.392959308" watchObservedRunningTime="2025-10-08 14:03:37.492850359 +0000 UTC m=+157.399597851" Oct 08 14:03:37 crc kubenswrapper[4789]: I1008 14:03:37.496660 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"12692816-f57d-4308-a946-261fee1ee271","Type":"ContainerDied","Data":"036a144be9a8c8aac0fbc95424d40aee9b9ae7feb8a0d57e4b79e79db4a32ad5"} Oct 08 14:03:37 crc kubenswrapper[4789]: I1008 14:03:37.496705 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="036a144be9a8c8aac0fbc95424d40aee9b9ae7feb8a0d57e4b79e79db4a32ad5" Oct 08 14:03:37 crc kubenswrapper[4789]: I1008 14:03:37.496770 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 14:03:38 crc kubenswrapper[4789]: I1008 14:03:38.120478 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-gr58t" Oct 08 14:03:38 crc kubenswrapper[4789]: I1008 14:03:38.510773 4789 generic.go:334] "Generic (PLEG): container finished" podID="9d8f6fac-c3df-4016-b978-ff0e17ea642f" containerID="d9e101585d9b4534e3297903c9578627ab41e00d5ef68dcabb8066288fece343" exitCode=0 Oct 08 14:03:38 crc kubenswrapper[4789]: I1008 14:03:38.510833 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9d8f6fac-c3df-4016-b978-ff0e17ea642f","Type":"ContainerDied","Data":"d9e101585d9b4534e3297903c9578627ab41e00d5ef68dcabb8066288fece343"} Oct 08 14:03:42 crc kubenswrapper[4789]: I1008 14:03:42.200058 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mvzv5" Oct 08 14:03:42 crc kubenswrapper[4789]: I1008 14:03:42.270038 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:42 crc kubenswrapper[4789]: I1008 14:03:42.273109 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:03:42 crc kubenswrapper[4789]: I1008 14:03:42.901947 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:42 crc kubenswrapper[4789]: I1008 14:03:42.907487 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765705a4-a303-4281-9677-5d0769f6d157-metrics-certs\") pod \"network-metrics-daemon-b269s\" (UID: \"765705a4-a303-4281-9677-5d0769f6d157\") " pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:42 crc kubenswrapper[4789]: I1008 14:03:42.947004 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b269s" Oct 08 14:03:48 crc kubenswrapper[4789]: I1008 14:03:48.654327 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:48 crc kubenswrapper[4789]: I1008 14:03:48.791509 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kube-api-access\") pod \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\" (UID: \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\") " Oct 08 14:03:48 crc kubenswrapper[4789]: I1008 14:03:48.791593 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kubelet-dir\") pod \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\" (UID: \"9d8f6fac-c3df-4016-b978-ff0e17ea642f\") " Oct 08 14:03:48 crc kubenswrapper[4789]: I1008 14:03:48.791758 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9d8f6fac-c3df-4016-b978-ff0e17ea642f" (UID: "9d8f6fac-c3df-4016-b978-ff0e17ea642f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:03:48 crc kubenswrapper[4789]: I1008 14:03:48.796171 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9d8f6fac-c3df-4016-b978-ff0e17ea642f" (UID: "9d8f6fac-c3df-4016-b978-ff0e17ea642f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:03:48 crc kubenswrapper[4789]: I1008 14:03:48.892892 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 14:03:48 crc kubenswrapper[4789]: I1008 14:03:48.892924 4789 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9d8f6fac-c3df-4016-b978-ff0e17ea642f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 14:03:49 crc kubenswrapper[4789]: I1008 14:03:49.249837 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:03:49 crc kubenswrapper[4789]: I1008 14:03:49.574751 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9d8f6fac-c3df-4016-b978-ff0e17ea642f","Type":"ContainerDied","Data":"338ce4bbdfc997917e3ce4dfe6b4fa41c4ed594be8536e01a7cb69fd36c2c817"} Oct 08 14:03:49 crc kubenswrapper[4789]: I1008 14:03:49.574794 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="338ce4bbdfc997917e3ce4dfe6b4fa41c4ed594be8536e01a7cb69fd36c2c817" Oct 08 14:03:49 crc kubenswrapper[4789]: I1008 14:03:49.574813 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 14:03:56 crc kubenswrapper[4789]: I1008 14:03:56.433067 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:03:56 crc kubenswrapper[4789]: I1008 14:03:56.433548 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:03:58 crc kubenswrapper[4789]: E1008 14:03:58.712080 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 14:03:58 crc kubenswrapper[4789]: E1008 14:03:58.712622 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7cw5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lzcr8_openshift-marketplace(a3054fb2-f238-47a2-9cc0-8ae3fb42d65a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:03:58 crc kubenswrapper[4789]: E1008 14:03:58.713839 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lzcr8" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" Oct 08 14:03:58 crc kubenswrapper[4789]: E1008 14:03:58.718132 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 14:03:58 crc kubenswrapper[4789]: E1008 14:03:58.718321 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d59rc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-z84nj_openshift-marketplace(35d2ebf2-6bd4-4adf-872d-d241a2519b0d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:03:58 crc kubenswrapper[4789]: E1008 14:03:58.719714 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-z84nj" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" Oct 08 14:04:02 crc kubenswrapper[4789]: I1008 14:04:02.466128 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hzztl" Oct 08 14:04:04 crc kubenswrapper[4789]: E1008 14:04:04.020475 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-z84nj" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" Oct 08 14:04:04 crc kubenswrapper[4789]: E1008 14:04:04.020900 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lzcr8" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" Oct 08 14:04:05 crc kubenswrapper[4789]: E1008 14:04:05.782079 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 14:04:05 crc kubenswrapper[4789]: E1008 14:04:05.782254 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gjbkg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vn2fk_openshift-marketplace(68b331a3-a0a4-4fc8-bcd0-7d180f2a86be): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:04:05 crc kubenswrapper[4789]: E1008 14:04:05.783646 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-vn2fk" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" Oct 08 14:04:08 crc kubenswrapper[4789]: E1008 14:04:08.452123 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vn2fk" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" Oct 08 14:04:08 crc kubenswrapper[4789]: E1008 14:04:08.554408 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 14:04:08 crc kubenswrapper[4789]: E1008 14:04:08.554879 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-69nvc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-p9cvl_openshift-marketplace(dc0408a7-643a-4260-9d54-382cdb73caca): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:04:08 crc kubenswrapper[4789]: E1008 14:04:08.556389 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-p9cvl" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.645544 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-p9cvl" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.724835 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.725392 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2676f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rqnlm_openshift-marketplace(5e523982-53d8-4082-a367-d99a110d544f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.726526 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rqnlm" podUID="5e523982-53d8-4082-a367-d99a110d544f" Oct 08 14:04:09 crc kubenswrapper[4789]: I1008 14:04:09.778716 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.781090 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.781235 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bb6sq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9z9p6_openshift-marketplace(95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.782437 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-9z9p6" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.800874 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.801066 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5dwfh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-72nk5_openshift-marketplace(a5e65840-1d44-4cb3-bd42-3bc0048f7274): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.802266 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-72nk5" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.806943 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.807202 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2rbws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rswj8_openshift-marketplace(2908d4c0-bbe0-436b-b806-ca29cf20afcf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:04:09 crc kubenswrapper[4789]: E1008 14:04:09.808404 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rswj8" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" Oct 08 14:04:10 crc kubenswrapper[4789]: I1008 14:04:10.045189 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b269s"] Oct 08 14:04:10 crc kubenswrapper[4789]: W1008 14:04:10.050105 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod765705a4_a303_4281_9677_5d0769f6d157.slice/crio-a8fdb54e966838ac2f990a1373ed1cce42020bd2169fc38f36ff71405f3a18ba WatchSource:0}: Error finding container a8fdb54e966838ac2f990a1373ed1cce42020bd2169fc38f36ff71405f3a18ba: Status 404 returned error can't find the container with id a8fdb54e966838ac2f990a1373ed1cce42020bd2169fc38f36ff71405f3a18ba Oct 08 14:04:10 crc kubenswrapper[4789]: I1008 14:04:10.684517 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b269s" event={"ID":"765705a4-a303-4281-9677-5d0769f6d157","Type":"ContainerStarted","Data":"4056a21f419f4f6fbe1555fe587e893dd6e6b70255aa1ef958491b2a765ab6e3"} Oct 08 14:04:10 crc kubenswrapper[4789]: I1008 14:04:10.684742 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b269s" event={"ID":"765705a4-a303-4281-9677-5d0769f6d157","Type":"ContainerStarted","Data":"fcc5a355131e74c46456bd8f51ddada9a80bfea69e15e2e8180e456cc909f5a2"} Oct 08 14:04:10 crc kubenswrapper[4789]: I1008 14:04:10.684757 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b269s" event={"ID":"765705a4-a303-4281-9677-5d0769f6d157","Type":"ContainerStarted","Data":"a8fdb54e966838ac2f990a1373ed1cce42020bd2169fc38f36ff71405f3a18ba"} Oct 08 14:04:10 crc kubenswrapper[4789]: E1008 14:04:10.686788 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rswj8" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" Oct 08 14:04:10 crc kubenswrapper[4789]: E1008 14:04:10.686800 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rqnlm" podUID="5e523982-53d8-4082-a367-d99a110d544f" Oct 08 14:04:10 crc kubenswrapper[4789]: E1008 14:04:10.686800 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-72nk5" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" Oct 08 14:04:10 crc kubenswrapper[4789]: E1008 14:04:10.686922 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9z9p6" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" Oct 08 14:04:10 crc kubenswrapper[4789]: I1008 14:04:10.716342 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-b269s" podStartSLOduration=169.716327287 podStartE2EDuration="2m49.716327287s" podCreationTimestamp="2025-10-08 14:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:04:10.713947841 +0000 UTC m=+190.620695353" watchObservedRunningTime="2025-10-08 14:04:10.716327287 +0000 UTC m=+190.623074779" Oct 08 14:04:16 crc kubenswrapper[4789]: I1008 14:04:16.719662 4789 generic.go:334] "Generic (PLEG): container finished" podID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerID="e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337" exitCode=0 Oct 08 14:04:16 crc kubenswrapper[4789]: I1008 14:04:16.719702 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z84nj" event={"ID":"35d2ebf2-6bd4-4adf-872d-d241a2519b0d","Type":"ContainerDied","Data":"e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337"} Oct 08 14:04:16 crc kubenswrapper[4789]: I1008 14:04:16.727426 4789 generic.go:334] "Generic (PLEG): container finished" podID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerID="d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020" exitCode=0 Oct 08 14:04:16 crc kubenswrapper[4789]: I1008 14:04:16.727474 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcr8" event={"ID":"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a","Type":"ContainerDied","Data":"d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020"} Oct 08 14:04:17 crc kubenswrapper[4789]: I1008 14:04:17.739418 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z84nj" event={"ID":"35d2ebf2-6bd4-4adf-872d-d241a2519b0d","Type":"ContainerStarted","Data":"48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6"} Oct 08 14:04:17 crc kubenswrapper[4789]: I1008 14:04:17.744921 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcr8" event={"ID":"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a","Type":"ContainerStarted","Data":"752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f"} Oct 08 14:04:17 crc kubenswrapper[4789]: I1008 14:04:17.786755 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z84nj" podStartSLOduration=1.6852366509999999 podStartE2EDuration="46.786732643s" podCreationTimestamp="2025-10-08 14:03:31 +0000 UTC" firstStartedPulling="2025-10-08 14:03:32.232728029 +0000 UTC m=+152.139475521" lastFinishedPulling="2025-10-08 14:04:17.334224011 +0000 UTC m=+197.240971513" observedRunningTime="2025-10-08 14:04:17.78407169 +0000 UTC m=+197.690819212" watchObservedRunningTime="2025-10-08 14:04:17.786732643 +0000 UTC m=+197.693480155" Oct 08 14:04:21 crc kubenswrapper[4789]: I1008 14:04:21.732181 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:04:21 crc kubenswrapper[4789]: I1008 14:04:21.732888 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:04:21 crc kubenswrapper[4789]: I1008 14:04:21.759105 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lzcr8" podStartSLOduration=6.905657889 podStartE2EDuration="50.759072387s" podCreationTimestamp="2025-10-08 14:03:31 +0000 UTC" firstStartedPulling="2025-10-08 14:03:33.337482436 +0000 UTC m=+153.244229928" lastFinishedPulling="2025-10-08 14:04:17.190896904 +0000 UTC m=+197.097644426" observedRunningTime="2025-10-08 14:04:17.818139898 +0000 UTC m=+197.724887400" watchObservedRunningTime="2025-10-08 14:04:21.759072387 +0000 UTC m=+201.665819919" Oct 08 14:04:21 crc kubenswrapper[4789]: I1008 14:04:21.901577 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:04:22 crc kubenswrapper[4789]: I1008 14:04:22.116881 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:04:22 crc kubenswrapper[4789]: I1008 14:04:22.117050 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:04:22 crc kubenswrapper[4789]: I1008 14:04:22.164215 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:04:22 crc kubenswrapper[4789]: I1008 14:04:22.833369 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:04:22 crc kubenswrapper[4789]: I1008 14:04:22.833852 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:04:23 crc kubenswrapper[4789]: I1008 14:04:23.779386 4789 generic.go:334] "Generic (PLEG): container finished" podID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerID="128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3" exitCode=0 Oct 08 14:04:23 crc kubenswrapper[4789]: I1008 14:04:23.779528 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72nk5" event={"ID":"a5e65840-1d44-4cb3-bd42-3bc0048f7274","Type":"ContainerDied","Data":"128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3"} Oct 08 14:04:23 crc kubenswrapper[4789]: I1008 14:04:23.786663 4789 generic.go:334] "Generic (PLEG): container finished" podID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerID="5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585" exitCode=0 Oct 08 14:04:23 crc kubenswrapper[4789]: I1008 14:04:23.786785 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2fk" event={"ID":"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be","Type":"ContainerDied","Data":"5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585"} Oct 08 14:04:24 crc kubenswrapper[4789]: I1008 14:04:24.794355 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2fk" event={"ID":"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be","Type":"ContainerStarted","Data":"15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7"} Oct 08 14:04:24 crc kubenswrapper[4789]: I1008 14:04:24.796901 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72nk5" event={"ID":"a5e65840-1d44-4cb3-bd42-3bc0048f7274","Type":"ContainerStarted","Data":"04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6"} Oct 08 14:04:24 crc kubenswrapper[4789]: I1008 14:04:24.798760 4789 generic.go:334] "Generic (PLEG): container finished" podID="dc0408a7-643a-4260-9d54-382cdb73caca" containerID="78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897" exitCode=0 Oct 08 14:04:24 crc kubenswrapper[4789]: I1008 14:04:24.798815 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9cvl" event={"ID":"dc0408a7-643a-4260-9d54-382cdb73caca","Type":"ContainerDied","Data":"78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897"} Oct 08 14:04:24 crc kubenswrapper[4789]: I1008 14:04:24.800729 4789 generic.go:334] "Generic (PLEG): container finished" podID="5e523982-53d8-4082-a367-d99a110d544f" containerID="183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701" exitCode=0 Oct 08 14:04:24 crc kubenswrapper[4789]: I1008 14:04:24.800813 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnlm" event={"ID":"5e523982-53d8-4082-a367-d99a110d544f","Type":"ContainerDied","Data":"183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701"} Oct 08 14:04:24 crc kubenswrapper[4789]: I1008 14:04:24.813692 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vn2fk" podStartSLOduration=2.857188298 podStartE2EDuration="52.813670337s" podCreationTimestamp="2025-10-08 14:03:32 +0000 UTC" firstStartedPulling="2025-10-08 14:03:34.386068264 +0000 UTC m=+154.292815756" lastFinishedPulling="2025-10-08 14:04:24.342550303 +0000 UTC m=+204.249297795" observedRunningTime="2025-10-08 14:04:24.80997946 +0000 UTC m=+204.716726952" watchObservedRunningTime="2025-10-08 14:04:24.813670337 +0000 UTC m=+204.720417829" Oct 08 14:04:24 crc kubenswrapper[4789]: I1008 14:04:24.846709 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-72nk5" podStartSLOduration=3.777824204 podStartE2EDuration="56.846691341s" podCreationTimestamp="2025-10-08 14:03:28 +0000 UTC" firstStartedPulling="2025-10-08 14:03:31.230154227 +0000 UTC m=+151.136901719" lastFinishedPulling="2025-10-08 14:04:24.299021354 +0000 UTC m=+204.205768856" observedRunningTime="2025-10-08 14:04:24.843584659 +0000 UTC m=+204.750332151" watchObservedRunningTime="2025-10-08 14:04:24.846691341 +0000 UTC m=+204.753438833" Oct 08 14:04:25 crc kubenswrapper[4789]: I1008 14:04:25.165327 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcr8"] Oct 08 14:04:25 crc kubenswrapper[4789]: I1008 14:04:25.807483 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9cvl" event={"ID":"dc0408a7-643a-4260-9d54-382cdb73caca","Type":"ContainerStarted","Data":"d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581"} Oct 08 14:04:25 crc kubenswrapper[4789]: I1008 14:04:25.809328 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnlm" event={"ID":"5e523982-53d8-4082-a367-d99a110d544f","Type":"ContainerStarted","Data":"c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4"} Oct 08 14:04:25 crc kubenswrapper[4789]: I1008 14:04:25.810867 4789 generic.go:334] "Generic (PLEG): container finished" podID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerID="ca61a81b00ec13d9475967dbc7834d5e9e4099fadf87d4117dcef52770e2063a" exitCode=0 Oct 08 14:04:25 crc kubenswrapper[4789]: I1008 14:04:25.810955 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z9p6" event={"ID":"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6","Type":"ContainerDied","Data":"ca61a81b00ec13d9475967dbc7834d5e9e4099fadf87d4117dcef52770e2063a"} Oct 08 14:04:25 crc kubenswrapper[4789]: I1008 14:04:25.811094 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lzcr8" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerName="registry-server" containerID="cri-o://752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f" gracePeriod=2 Oct 08 14:04:25 crc kubenswrapper[4789]: I1008 14:04:25.830057 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p9cvl" podStartSLOduration=2.653842927 podStartE2EDuration="56.830035276s" podCreationTimestamp="2025-10-08 14:03:29 +0000 UTC" firstStartedPulling="2025-10-08 14:03:31.205047726 +0000 UTC m=+151.111795218" lastFinishedPulling="2025-10-08 14:04:25.381240075 +0000 UTC m=+205.287987567" observedRunningTime="2025-10-08 14:04:25.828612328 +0000 UTC m=+205.735359840" watchObservedRunningTime="2025-10-08 14:04:25.830035276 +0000 UTC m=+205.736782788" Oct 08 14:04:25 crc kubenswrapper[4789]: I1008 14:04:25.865511 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rqnlm" podStartSLOduration=2.73926485 podStartE2EDuration="56.865492293s" podCreationTimestamp="2025-10-08 14:03:29 +0000 UTC" firstStartedPulling="2025-10-08 14:03:31.211542715 +0000 UTC m=+151.118290207" lastFinishedPulling="2025-10-08 14:04:25.337770158 +0000 UTC m=+205.244517650" observedRunningTime="2025-10-08 14:04:25.864018405 +0000 UTC m=+205.770765897" watchObservedRunningTime="2025-10-08 14:04:25.865492293 +0000 UTC m=+205.772239785" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.216840 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.406024 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cw5x\" (UniqueName: \"kubernetes.io/projected/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-kube-api-access-7cw5x\") pod \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.406405 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-catalog-content\") pod \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.406440 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-utilities\") pod \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\" (UID: \"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a\") " Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.407174 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-utilities" (OuterVolumeSpecName: "utilities") pod "a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" (UID: "a3054fb2-f238-47a2-9cc0-8ae3fb42d65a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.420682 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" (UID: "a3054fb2-f238-47a2-9cc0-8ae3fb42d65a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.420846 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-kube-api-access-7cw5x" (OuterVolumeSpecName: "kube-api-access-7cw5x") pod "a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" (UID: "a3054fb2-f238-47a2-9cc0-8ae3fb42d65a"). InnerVolumeSpecName "kube-api-access-7cw5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.433223 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.433310 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.433383 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.434141 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.434305 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f" gracePeriod=600 Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.508325 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.508621 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.508711 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cw5x\" (UniqueName: \"kubernetes.io/projected/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a-kube-api-access-7cw5x\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.816195 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f" exitCode=0 Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.816277 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f"} Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.818497 4789 generic.go:334] "Generic (PLEG): container finished" podID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerID="752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f" exitCode=0 Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.818530 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcr8" event={"ID":"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a","Type":"ContainerDied","Data":"752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f"} Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.818552 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcr8" event={"ID":"a3054fb2-f238-47a2-9cc0-8ae3fb42d65a","Type":"ContainerDied","Data":"2058556dfcd0f84ddccc8815917bfc4d03ed8caed386fa6f59feecf4b6fb35c4"} Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.818574 4789 scope.go:117] "RemoveContainer" containerID="752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.818699 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzcr8" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.836119 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcr8"] Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.836279 4789 scope.go:117] "RemoveContainer" containerID="d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.839770 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcr8"] Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.851029 4789 scope.go:117] "RemoveContainer" containerID="f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.865963 4789 scope.go:117] "RemoveContainer" containerID="752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f" Oct 08 14:04:26 crc kubenswrapper[4789]: E1008 14:04:26.866332 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f\": container with ID starting with 752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f not found: ID does not exist" containerID="752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.866370 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f"} err="failed to get container status \"752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f\": rpc error: code = NotFound desc = could not find container \"752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f\": container with ID starting with 752c1373b72171a39291ce267d127784ce581f2918ec0d8d07d7076b8901442f not found: ID does not exist" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.866398 4789 scope.go:117] "RemoveContainer" containerID="d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020" Oct 08 14:04:26 crc kubenswrapper[4789]: E1008 14:04:26.866709 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020\": container with ID starting with d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020 not found: ID does not exist" containerID="d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.866754 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020"} err="failed to get container status \"d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020\": rpc error: code = NotFound desc = could not find container \"d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020\": container with ID starting with d67f7c965c5092705d7c479bf0a33c4c4f38d83d9fdb3fd43aa599b07a7b0020 not found: ID does not exist" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.866785 4789 scope.go:117] "RemoveContainer" containerID="f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b" Oct 08 14:04:26 crc kubenswrapper[4789]: E1008 14:04:26.867085 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b\": container with ID starting with f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b not found: ID does not exist" containerID="f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b" Oct 08 14:04:26 crc kubenswrapper[4789]: I1008 14:04:26.867112 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b"} err="failed to get container status \"f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b\": rpc error: code = NotFound desc = could not find container \"f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b\": container with ID starting with f3d5a3cdb8e84f40cd64a79e73336313ce6033bf2aa7720ba304c235924f665b not found: ID does not exist" Oct 08 14:04:27 crc kubenswrapper[4789]: I1008 14:04:27.827201 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"c481a44e4e2f4fdaae1b3ca23f477eaf282c238b6ce960c39dd22247b81d0297"} Oct 08 14:04:27 crc kubenswrapper[4789]: I1008 14:04:27.832797 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z9p6" event={"ID":"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6","Type":"ContainerStarted","Data":"bf1cbaa088b6c0d377f78f80468153c35215969619e3a432b58858fcae06bb71"} Oct 08 14:04:27 crc kubenswrapper[4789]: I1008 14:04:27.834722 4789 generic.go:334] "Generic (PLEG): container finished" podID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerID="7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49" exitCode=0 Oct 08 14:04:27 crc kubenswrapper[4789]: I1008 14:04:27.834770 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rswj8" event={"ID":"2908d4c0-bbe0-436b-b806-ca29cf20afcf","Type":"ContainerDied","Data":"7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49"} Oct 08 14:04:27 crc kubenswrapper[4789]: I1008 14:04:27.893910 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9z9p6" podStartSLOduration=4.195211073 podStartE2EDuration="55.893892968s" podCreationTimestamp="2025-10-08 14:03:32 +0000 UTC" firstStartedPulling="2025-10-08 14:03:35.414681844 +0000 UTC m=+155.321429336" lastFinishedPulling="2025-10-08 14:04:27.113363739 +0000 UTC m=+207.020111231" observedRunningTime="2025-10-08 14:04:27.891605588 +0000 UTC m=+207.798353090" watchObservedRunningTime="2025-10-08 14:04:27.893892968 +0000 UTC m=+207.800640470" Oct 08 14:04:28 crc kubenswrapper[4789]: I1008 14:04:28.736547 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" path="/var/lib/kubelet/pods/a3054fb2-f238-47a2-9cc0-8ae3fb42d65a/volumes" Oct 08 14:04:28 crc kubenswrapper[4789]: I1008 14:04:28.841251 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rswj8" event={"ID":"2908d4c0-bbe0-436b-b806-ca29cf20afcf","Type":"ContainerStarted","Data":"7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba"} Oct 08 14:04:28 crc kubenswrapper[4789]: I1008 14:04:28.861444 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rswj8" podStartSLOduration=2.560275421 podStartE2EDuration="59.861431098s" podCreationTimestamp="2025-10-08 14:03:29 +0000 UTC" firstStartedPulling="2025-10-08 14:03:31.214097735 +0000 UTC m=+151.120845227" lastFinishedPulling="2025-10-08 14:04:28.515253412 +0000 UTC m=+208.422000904" observedRunningTime="2025-10-08 14:04:28.858093341 +0000 UTC m=+208.764840833" watchObservedRunningTime="2025-10-08 14:04:28.861431098 +0000 UTC m=+208.768178590" Oct 08 14:04:29 crc kubenswrapper[4789]: I1008 14:04:29.323839 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:04:29 crc kubenswrapper[4789]: I1008 14:04:29.326094 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:04:29 crc kubenswrapper[4789]: I1008 14:04:29.386720 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:04:29 crc kubenswrapper[4789]: I1008 14:04:29.745249 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:04:29 crc kubenswrapper[4789]: I1008 14:04:29.745295 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:04:29 crc kubenswrapper[4789]: I1008 14:04:29.785341 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:04:29 crc kubenswrapper[4789]: I1008 14:04:29.882697 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:04:29 crc kubenswrapper[4789]: I1008 14:04:29.977403 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:04:29 crc kubenswrapper[4789]: I1008 14:04:29.977662 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:04:30 crc kubenswrapper[4789]: I1008 14:04:30.116320 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:04:30 crc kubenswrapper[4789]: I1008 14:04:30.116375 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:04:30 crc kubenswrapper[4789]: I1008 14:04:30.152666 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:04:30 crc kubenswrapper[4789]: I1008 14:04:30.890650 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:04:31 crc kubenswrapper[4789]: I1008 14:04:31.015790 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rswj8" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerName="registry-server" probeResult="failure" output=< Oct 08 14:04:31 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 14:04:31 crc kubenswrapper[4789]: > Oct 08 14:04:32 crc kubenswrapper[4789]: I1008 14:04:32.842780 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:04:32 crc kubenswrapper[4789]: I1008 14:04:32.843378 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:04:32 crc kubenswrapper[4789]: I1008 14:04:32.886815 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:04:32 crc kubenswrapper[4789]: I1008 14:04:32.927775 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:04:32 crc kubenswrapper[4789]: I1008 14:04:32.965016 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9cvl"] Oct 08 14:04:32 crc kubenswrapper[4789]: I1008 14:04:32.965300 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p9cvl" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" containerName="registry-server" containerID="cri-o://d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581" gracePeriod=2 Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.154662 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.155245 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.201177 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.345839 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.392717 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69nvc\" (UniqueName: \"kubernetes.io/projected/dc0408a7-643a-4260-9d54-382cdb73caca-kube-api-access-69nvc\") pod \"dc0408a7-643a-4260-9d54-382cdb73caca\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.392831 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-utilities\") pod \"dc0408a7-643a-4260-9d54-382cdb73caca\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.393762 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-utilities" (OuterVolumeSpecName: "utilities") pod "dc0408a7-643a-4260-9d54-382cdb73caca" (UID: "dc0408a7-643a-4260-9d54-382cdb73caca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.393845 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-catalog-content\") pod \"dc0408a7-643a-4260-9d54-382cdb73caca\" (UID: \"dc0408a7-643a-4260-9d54-382cdb73caca\") " Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.394627 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.400677 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc0408a7-643a-4260-9d54-382cdb73caca-kube-api-access-69nvc" (OuterVolumeSpecName: "kube-api-access-69nvc") pod "dc0408a7-643a-4260-9d54-382cdb73caca" (UID: "dc0408a7-643a-4260-9d54-382cdb73caca"). InnerVolumeSpecName "kube-api-access-69nvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.448908 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc0408a7-643a-4260-9d54-382cdb73caca" (UID: "dc0408a7-643a-4260-9d54-382cdb73caca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.495751 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69nvc\" (UniqueName: \"kubernetes.io/projected/dc0408a7-643a-4260-9d54-382cdb73caca-kube-api-access-69nvc\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.495796 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0408a7-643a-4260-9d54-382cdb73caca-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.867834 4789 generic.go:334] "Generic (PLEG): container finished" podID="dc0408a7-643a-4260-9d54-382cdb73caca" containerID="d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581" exitCode=0 Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.867929 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9cvl" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.867920 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9cvl" event={"ID":"dc0408a7-643a-4260-9d54-382cdb73caca","Type":"ContainerDied","Data":"d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581"} Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.868021 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9cvl" event={"ID":"dc0408a7-643a-4260-9d54-382cdb73caca","Type":"ContainerDied","Data":"2f9fa3ab463bb90792a3c8f602dc7280982fde3c984fc4b7f025758759a1fa8e"} Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.868054 4789 scope.go:117] "RemoveContainer" containerID="d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.890602 4789 scope.go:117] "RemoveContainer" containerID="78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.901922 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9cvl"] Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.905208 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p9cvl"] Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.918214 4789 scope.go:117] "RemoveContainer" containerID="7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.920131 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.934526 4789 scope.go:117] "RemoveContainer" containerID="d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581" Oct 08 14:04:33 crc kubenswrapper[4789]: E1008 14:04:33.935000 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581\": container with ID starting with d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581 not found: ID does not exist" containerID="d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.935049 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581"} err="failed to get container status \"d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581\": rpc error: code = NotFound desc = could not find container \"d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581\": container with ID starting with d5e494b4648cb47dfd86ad0ad224fe330b6d9cf024bbac2791390215d91fc581 not found: ID does not exist" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.935079 4789 scope.go:117] "RemoveContainer" containerID="78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897" Oct 08 14:04:33 crc kubenswrapper[4789]: E1008 14:04:33.935558 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897\": container with ID starting with 78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897 not found: ID does not exist" containerID="78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.935609 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897"} err="failed to get container status \"78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897\": rpc error: code = NotFound desc = could not find container \"78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897\": container with ID starting with 78e14566554f27273b891b1660c0aab2919ccda733b018ec475064150a40c897 not found: ID does not exist" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.935643 4789 scope.go:117] "RemoveContainer" containerID="7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb" Oct 08 14:04:33 crc kubenswrapper[4789]: E1008 14:04:33.936030 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb\": container with ID starting with 7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb not found: ID does not exist" containerID="7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb" Oct 08 14:04:33 crc kubenswrapper[4789]: I1008 14:04:33.936090 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb"} err="failed to get container status \"7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb\": rpc error: code = NotFound desc = could not find container \"7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb\": container with ID starting with 7544a141f8fadf815cbf46bd6c31eb9320a607027ecd8742afcb14cbc72c2cfb not found: ID does not exist" Oct 08 14:04:34 crc kubenswrapper[4789]: I1008 14:04:34.738221 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" path="/var/lib/kubelet/pods/dc0408a7-643a-4260-9d54-382cdb73caca/volumes" Oct 08 14:04:35 crc kubenswrapper[4789]: I1008 14:04:35.565704 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9z9p6"] Oct 08 14:04:35 crc kubenswrapper[4789]: I1008 14:04:35.879796 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9z9p6" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerName="registry-server" containerID="cri-o://bf1cbaa088b6c0d377f78f80468153c35215969619e3a432b58858fcae06bb71" gracePeriod=2 Oct 08 14:04:37 crc kubenswrapper[4789]: I1008 14:04:37.889004 4789 generic.go:334] "Generic (PLEG): container finished" podID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerID="bf1cbaa088b6c0d377f78f80468153c35215969619e3a432b58858fcae06bb71" exitCode=0 Oct 08 14:04:37 crc kubenswrapper[4789]: I1008 14:04:37.889098 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z9p6" event={"ID":"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6","Type":"ContainerDied","Data":"bf1cbaa088b6c0d377f78f80468153c35215969619e3a432b58858fcae06bb71"} Oct 08 14:04:37 crc kubenswrapper[4789]: I1008 14:04:37.965016 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.164117 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb6sq\" (UniqueName: \"kubernetes.io/projected/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-kube-api-access-bb6sq\") pod \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.164186 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-utilities\") pod \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.164205 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-catalog-content\") pod \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\" (UID: \"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6\") " Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.165245 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-utilities" (OuterVolumeSpecName: "utilities") pod "95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" (UID: "95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.169799 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-kube-api-access-bb6sq" (OuterVolumeSpecName: "kube-api-access-bb6sq") pod "95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" (UID: "95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6"). InnerVolumeSpecName "kube-api-access-bb6sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.266018 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb6sq\" (UniqueName: \"kubernetes.io/projected/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-kube-api-access-bb6sq\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.266046 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.295500 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" (UID: "95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.367609 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.895422 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z9p6" event={"ID":"95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6","Type":"ContainerDied","Data":"70a3c79d6120b673d9c20bfceddbd038e17aa087df781d5a555c43f71e7c8e95"} Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.895711 4789 scope.go:117] "RemoveContainer" containerID="bf1cbaa088b6c0d377f78f80468153c35215969619e3a432b58858fcae06bb71" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.895461 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9z9p6" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.908738 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9z9p6"] Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.911828 4789 scope.go:117] "RemoveContainer" containerID="ca61a81b00ec13d9475967dbc7834d5e9e4099fadf87d4117dcef52770e2063a" Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.912906 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9z9p6"] Oct 08 14:04:38 crc kubenswrapper[4789]: I1008 14:04:38.922785 4789 scope.go:117] "RemoveContainer" containerID="67ba6eb3b1cd3dc9e069f54a3ddbc91bda15455a07b6ee0ed59b896cab1508f2" Oct 08 14:04:39 crc kubenswrapper[4789]: I1008 14:04:39.784411 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:04:40 crc kubenswrapper[4789]: I1008 14:04:40.021662 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:04:40 crc kubenswrapper[4789]: I1008 14:04:40.060620 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:04:40 crc kubenswrapper[4789]: I1008 14:04:40.754292 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" path="/var/lib/kubelet/pods/95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6/volumes" Oct 08 14:04:41 crc kubenswrapper[4789]: I1008 14:04:41.245595 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mtgx9"] Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.167471 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqnlm"] Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.167946 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rqnlm" podUID="5e523982-53d8-4082-a367-d99a110d544f" containerName="registry-server" containerID="cri-o://c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4" gracePeriod=2 Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.550197 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.562875 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-utilities\") pod \"5e523982-53d8-4082-a367-d99a110d544f\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.562955 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-catalog-content\") pod \"5e523982-53d8-4082-a367-d99a110d544f\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.563014 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2676f\" (UniqueName: \"kubernetes.io/projected/5e523982-53d8-4082-a367-d99a110d544f-kube-api-access-2676f\") pod \"5e523982-53d8-4082-a367-d99a110d544f\" (UID: \"5e523982-53d8-4082-a367-d99a110d544f\") " Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.563928 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-utilities" (OuterVolumeSpecName: "utilities") pod "5e523982-53d8-4082-a367-d99a110d544f" (UID: "5e523982-53d8-4082-a367-d99a110d544f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.572959 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e523982-53d8-4082-a367-d99a110d544f-kube-api-access-2676f" (OuterVolumeSpecName: "kube-api-access-2676f") pod "5e523982-53d8-4082-a367-d99a110d544f" (UID: "5e523982-53d8-4082-a367-d99a110d544f"). InnerVolumeSpecName "kube-api-access-2676f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.609573 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e523982-53d8-4082-a367-d99a110d544f" (UID: "5e523982-53d8-4082-a367-d99a110d544f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.663844 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.663871 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e523982-53d8-4082-a367-d99a110d544f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.663881 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2676f\" (UniqueName: \"kubernetes.io/projected/5e523982-53d8-4082-a367-d99a110d544f-kube-api-access-2676f\") on node \"crc\" DevicePath \"\"" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.915474 4789 generic.go:334] "Generic (PLEG): container finished" podID="5e523982-53d8-4082-a367-d99a110d544f" containerID="c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4" exitCode=0 Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.915677 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnlm" event={"ID":"5e523982-53d8-4082-a367-d99a110d544f","Type":"ContainerDied","Data":"c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4"} Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.915723 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnlm" event={"ID":"5e523982-53d8-4082-a367-d99a110d544f","Type":"ContainerDied","Data":"e34663c74d429a97bfeacee83c4651a3d6bd72d8ef43a024af49351ffb2e356c"} Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.915741 4789 scope.go:117] "RemoveContainer" containerID="c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.915746 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqnlm" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.929784 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqnlm"] Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.931222 4789 scope.go:117] "RemoveContainer" containerID="183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.933742 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rqnlm"] Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.943658 4789 scope.go:117] "RemoveContainer" containerID="0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.960039 4789 scope.go:117] "RemoveContainer" containerID="c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4" Oct 08 14:04:42 crc kubenswrapper[4789]: E1008 14:04:42.960373 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4\": container with ID starting with c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4 not found: ID does not exist" containerID="c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.960416 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4"} err="failed to get container status \"c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4\": rpc error: code = NotFound desc = could not find container \"c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4\": container with ID starting with c13d8ab3e0f90df4c16dd90a55b148bb431024c333c4618aa10b41ab658876b4 not found: ID does not exist" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.960444 4789 scope.go:117] "RemoveContainer" containerID="183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701" Oct 08 14:04:42 crc kubenswrapper[4789]: E1008 14:04:42.960684 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701\": container with ID starting with 183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701 not found: ID does not exist" containerID="183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.960716 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701"} err="failed to get container status \"183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701\": rpc error: code = NotFound desc = could not find container \"183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701\": container with ID starting with 183a32e359f19b18e33c17b8ba86057b5917f6bcae87113d7995de03a38d4701 not found: ID does not exist" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.960739 4789 scope.go:117] "RemoveContainer" containerID="0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce" Oct 08 14:04:42 crc kubenswrapper[4789]: E1008 14:04:42.961018 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce\": container with ID starting with 0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce not found: ID does not exist" containerID="0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce" Oct 08 14:04:42 crc kubenswrapper[4789]: I1008 14:04:42.961045 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce"} err="failed to get container status \"0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce\": rpc error: code = NotFound desc = could not find container \"0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce\": container with ID starting with 0f366bd4d3a86ae65067cf0ef1369b795a6b57a4f4d08d3a0ccb1f8b72dfa9ce not found: ID does not exist" Oct 08 14:04:44 crc kubenswrapper[4789]: I1008 14:04:44.735655 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e523982-53d8-4082-a367-d99a110d544f" path="/var/lib/kubelet/pods/5e523982-53d8-4082-a367-d99a110d544f/volumes" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.278242 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" podUID="119eb4b6-56d0-4cae-803a-56020cda0096" containerName="oauth-openshift" containerID="cri-o://9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693" gracePeriod=15 Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.747957 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.783797 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-error\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.783846 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/119eb4b6-56d0-4cae-803a-56020cda0096-audit-dir\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.784035 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/119eb4b6-56d0-4cae-803a-56020cda0096-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.784232 4789 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/119eb4b6-56d0-4cae-803a-56020cda0096-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.790353 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.793859 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj"] Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794280 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e523982-53d8-4082-a367-d99a110d544f" containerName="extract-content" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794310 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e523982-53d8-4082-a367-d99a110d544f" containerName="extract-content" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794332 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12692816-f57d-4308-a946-261fee1ee271" containerName="pruner" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794345 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="12692816-f57d-4308-a946-261fee1ee271" containerName="pruner" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794365 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerName="extract-content" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794378 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerName="extract-content" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794396 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794407 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794427 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerName="extract-utilities" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794439 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerName="extract-utilities" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794454 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerName="extract-utilities" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794466 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerName="extract-utilities" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794482 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d8f6fac-c3df-4016-b978-ff0e17ea642f" containerName="pruner" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794493 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d8f6fac-c3df-4016-b978-ff0e17ea642f" containerName="pruner" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794509 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" containerName="extract-utilities" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794544 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" containerName="extract-utilities" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794564 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e523982-53d8-4082-a367-d99a110d544f" containerName="extract-utilities" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794576 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e523982-53d8-4082-a367-d99a110d544f" containerName="extract-utilities" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794594 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794606 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794626 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="119eb4b6-56d0-4cae-803a-56020cda0096" containerName="oauth-openshift" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794638 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="119eb4b6-56d0-4cae-803a-56020cda0096" containerName="oauth-openshift" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794656 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794668 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794684 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerName="extract-content" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794696 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerName="extract-content" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794722 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e523982-53d8-4082-a367-d99a110d544f" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794735 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e523982-53d8-4082-a367-d99a110d544f" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: E1008 14:05:06.794752 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" containerName="extract-content" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794764 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" containerName="extract-content" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794923 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d8f6fac-c3df-4016-b978-ff0e17ea642f" containerName="pruner" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794949 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="12692816-f57d-4308-a946-261fee1ee271" containerName="pruner" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.794978 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc0408a7-643a-4260-9d54-382cdb73caca" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.795036 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a38aa2-5ec1-442e-aaaf-2ccc7e4146f6" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.795065 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e523982-53d8-4082-a367-d99a110d544f" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.795091 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="119eb4b6-56d0-4cae-803a-56020cda0096" containerName="oauth-openshift" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.795109 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3054fb2-f238-47a2-9cc0-8ae3fb42d65a" containerName="registry-server" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.795682 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.802720 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj"] Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.884902 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-serving-cert\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.884960 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-ocp-branding-template\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885017 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-login\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885044 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-idp-0-file-data\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885079 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg7l8\" (UniqueName: \"kubernetes.io/projected/119eb4b6-56d0-4cae-803a-56020cda0096-kube-api-access-gg7l8\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885108 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-trusted-ca-bundle\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885133 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-audit-policies\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885234 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-cliconfig\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885292 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-session\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885320 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-router-certs\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885345 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-service-ca\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885368 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-provider-selection\") pod \"119eb4b6-56d0-4cae-803a-56020cda0096\" (UID: \"119eb4b6-56d0-4cae-803a-56020cda0096\") " Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885462 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885493 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885516 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-template-login\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885541 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885566 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/302b3c48-612f-43de-827c-b900f8a52f47-audit-dir\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885599 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-audit-policies\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885622 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-session\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885645 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885709 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885738 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-service-ca\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885771 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-router-certs\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885803 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g28h\" (UniqueName: \"kubernetes.io/projected/302b3c48-612f-43de-827c-b900f8a52f47-kube-api-access-5g28h\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885854 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-template-error\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885890 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885934 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885965 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.885959 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.886434 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.886610 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.890001 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.890817 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.890833 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/119eb4b6-56d0-4cae-803a-56020cda0096-kube-api-access-gg7l8" (OuterVolumeSpecName: "kube-api-access-gg7l8") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "kube-api-access-gg7l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.891143 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.891355 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.891417 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.891557 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.891734 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "119eb4b6-56d0-4cae-803a-56020cda0096" (UID: "119eb4b6-56d0-4cae-803a-56020cda0096"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.987654 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-audit-policies\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.987745 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-session\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.987805 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.987862 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.987901 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-service-ca\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.987940 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-router-certs\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988016 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g28h\" (UniqueName: \"kubernetes.io/projected/302b3c48-612f-43de-827c-b900f8a52f47-kube-api-access-5g28h\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988078 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-template-error\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988118 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988189 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988231 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988266 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-template-login\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988305 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988345 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/302b3c48-612f-43de-827c-b900f8a52f47-audit-dir\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988410 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988433 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988455 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988478 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988499 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988520 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988540 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988559 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988579 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg7l8\" (UniqueName: \"kubernetes.io/projected/119eb4b6-56d0-4cae-803a-56020cda0096-kube-api-access-gg7l8\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988599 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988621 4789 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988642 4789 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/119eb4b6-56d0-4cae-803a-56020cda0096-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988704 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/302b3c48-612f-43de-827c-b900f8a52f47-audit-dir\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.988915 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-service-ca\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.991346 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.991375 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-template-error\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.991593 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-audit-policies\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.992381 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-session\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.993342 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.993759 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.993872 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.994188 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-template-login\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.995123 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-router-certs\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.997066 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:06 crc kubenswrapper[4789]: I1008 14:05:06.997135 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/302b3c48-612f-43de-827c-b900f8a52f47-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.009887 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g28h\" (UniqueName: \"kubernetes.io/projected/302b3c48-612f-43de-827c-b900f8a52f47-kube-api-access-5g28h\") pod \"oauth-openshift-7557fdbcd4-z6dqj\" (UID: \"302b3c48-612f-43de-827c-b900f8a52f47\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.046347 4789 generic.go:334] "Generic (PLEG): container finished" podID="119eb4b6-56d0-4cae-803a-56020cda0096" containerID="9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693" exitCode=0 Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.046390 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" event={"ID":"119eb4b6-56d0-4cae-803a-56020cda0096","Type":"ContainerDied","Data":"9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693"} Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.046420 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" event={"ID":"119eb4b6-56d0-4cae-803a-56020cda0096","Type":"ContainerDied","Data":"ad085a17b1cdcc909bab079580c0bb0485e954897be3aeb2ed0f0f126ed44603"} Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.046434 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mtgx9" Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.046440 4789 scope.go:117] "RemoveContainer" containerID="9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693" Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.074279 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mtgx9"] Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.074463 4789 scope.go:117] "RemoveContainer" containerID="9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693" Oct 08 14:05:07 crc kubenswrapper[4789]: E1008 14:05:07.074866 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693\": container with ID starting with 9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693 not found: ID does not exist" containerID="9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693" Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.074902 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693"} err="failed to get container status \"9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693\": rpc error: code = NotFound desc = could not find container \"9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693\": container with ID starting with 9186f3bff8936933052bf863d15497747b036ff1ffcd6b1630e26f9094767693 not found: ID does not exist" Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.077368 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mtgx9"] Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.128367 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:07 crc kubenswrapper[4789]: I1008 14:05:07.536776 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj"] Oct 08 14:05:07 crc kubenswrapper[4789]: W1008 14:05:07.546263 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod302b3c48_612f_43de_827c_b900f8a52f47.slice/crio-b8a590b669cc01f0a816f8feade7976baf5e4ff284c87f2c96e5bb90c50e6a27 WatchSource:0}: Error finding container b8a590b669cc01f0a816f8feade7976baf5e4ff284c87f2c96e5bb90c50e6a27: Status 404 returned error can't find the container with id b8a590b669cc01f0a816f8feade7976baf5e4ff284c87f2c96e5bb90c50e6a27 Oct 08 14:05:08 crc kubenswrapper[4789]: I1008 14:05:08.060579 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" event={"ID":"302b3c48-612f-43de-827c-b900f8a52f47","Type":"ContainerStarted","Data":"908508d73013fccd67010050bc1ae87b749b44a2a30898fff7c28d5b0da14a48"} Oct 08 14:05:08 crc kubenswrapper[4789]: I1008 14:05:08.060624 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" event={"ID":"302b3c48-612f-43de-827c-b900f8a52f47","Type":"ContainerStarted","Data":"b8a590b669cc01f0a816f8feade7976baf5e4ff284c87f2c96e5bb90c50e6a27"} Oct 08 14:05:08 crc kubenswrapper[4789]: I1008 14:05:08.061199 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:08 crc kubenswrapper[4789]: I1008 14:05:08.117375 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" Oct 08 14:05:08 crc kubenswrapper[4789]: I1008 14:05:08.144507 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7557fdbcd4-z6dqj" podStartSLOduration=27.144486354 podStartE2EDuration="27.144486354s" podCreationTimestamp="2025-10-08 14:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:05:08.086040415 +0000 UTC m=+247.992787927" watchObservedRunningTime="2025-10-08 14:05:08.144486354 +0000 UTC m=+248.051233856" Oct 08 14:05:08 crc kubenswrapper[4789]: I1008 14:05:08.742390 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="119eb4b6-56d0-4cae-803a-56020cda0096" path="/var/lib/kubelet/pods/119eb4b6-56d0-4cae-803a-56020cda0096/volumes" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.144250 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rswj8"] Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.147843 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72nk5"] Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.148384 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-72nk5" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerName="registry-server" containerID="cri-o://04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6" gracePeriod=30 Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.148646 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rswj8" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerName="registry-server" containerID="cri-o://7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba" gracePeriod=30 Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.149180 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vzq6g"] Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.149408 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" podUID="5b3f76e2-8cce-47f1-87e5-43be0d07238e" containerName="marketplace-operator" containerID="cri-o://dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf" gracePeriod=30 Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.163538 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z84nj"] Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.163783 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z84nj" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerName="registry-server" containerID="cri-o://48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6" gracePeriod=30 Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.171622 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vn2fk"] Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.171843 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vn2fk" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerName="registry-server" containerID="cri-o://15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7" gracePeriod=30 Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.181230 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4lxzv"] Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.182103 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.195096 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4lxzv"] Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.228527 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c59c8d39-3840-48cf-b510-4e07d9293fc9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4lxzv\" (UID: \"c59c8d39-3840-48cf-b510-4e07d9293fc9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.228609 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjrr9\" (UniqueName: \"kubernetes.io/projected/c59c8d39-3840-48cf-b510-4e07d9293fc9-kube-api-access-zjrr9\") pod \"marketplace-operator-79b997595-4lxzv\" (UID: \"c59c8d39-3840-48cf-b510-4e07d9293fc9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.228674 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c59c8d39-3840-48cf-b510-4e07d9293fc9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4lxzv\" (UID: \"c59c8d39-3840-48cf-b510-4e07d9293fc9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.330755 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c59c8d39-3840-48cf-b510-4e07d9293fc9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4lxzv\" (UID: \"c59c8d39-3840-48cf-b510-4e07d9293fc9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.331355 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjrr9\" (UniqueName: \"kubernetes.io/projected/c59c8d39-3840-48cf-b510-4e07d9293fc9-kube-api-access-zjrr9\") pod \"marketplace-operator-79b997595-4lxzv\" (UID: \"c59c8d39-3840-48cf-b510-4e07d9293fc9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.331394 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c59c8d39-3840-48cf-b510-4e07d9293fc9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4lxzv\" (UID: \"c59c8d39-3840-48cf-b510-4e07d9293fc9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.333396 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c59c8d39-3840-48cf-b510-4e07d9293fc9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4lxzv\" (UID: \"c59c8d39-3840-48cf-b510-4e07d9293fc9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.338981 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c59c8d39-3840-48cf-b510-4e07d9293fc9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4lxzv\" (UID: \"c59c8d39-3840-48cf-b510-4e07d9293fc9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.358018 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjrr9\" (UniqueName: \"kubernetes.io/projected/c59c8d39-3840-48cf-b510-4e07d9293fc9-kube-api-access-zjrr9\") pod \"marketplace-operator-79b997595-4lxzv\" (UID: \"c59c8d39-3840-48cf-b510-4e07d9293fc9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.507741 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.624160 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.698412 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.702168 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.707101 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.710951 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.739029 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-operator-metrics\") pod \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.739275 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h9lt\" (UniqueName: \"kubernetes.io/projected/5b3f76e2-8cce-47f1-87e5-43be0d07238e-kube-api-access-2h9lt\") pod \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.739312 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-utilities\") pod \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.739380 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-utilities\") pod \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.739403 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-trusted-ca\") pod \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\" (UID: \"5b3f76e2-8cce-47f1-87e5-43be0d07238e\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.739423 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjbkg\" (UniqueName: \"kubernetes.io/projected/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-kube-api-access-gjbkg\") pod \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.741194 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-utilities" (OuterVolumeSpecName: "utilities") pod "2908d4c0-bbe0-436b-b806-ca29cf20afcf" (UID: "2908d4c0-bbe0-436b-b806-ca29cf20afcf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.742088 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-utilities" (OuterVolumeSpecName: "utilities") pod "a5e65840-1d44-4cb3-bd42-3bc0048f7274" (UID: "a5e65840-1d44-4cb3-bd42-3bc0048f7274"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.746259 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "5b3f76e2-8cce-47f1-87e5-43be0d07238e" (UID: "5b3f76e2-8cce-47f1-87e5-43be0d07238e"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.747502 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "5b3f76e2-8cce-47f1-87e5-43be0d07238e" (UID: "5b3f76e2-8cce-47f1-87e5-43be0d07238e"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.753459 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b3f76e2-8cce-47f1-87e5-43be0d07238e-kube-api-access-2h9lt" (OuterVolumeSpecName: "kube-api-access-2h9lt") pod "5b3f76e2-8cce-47f1-87e5-43be0d07238e" (UID: "5b3f76e2-8cce-47f1-87e5-43be0d07238e"). InnerVolumeSpecName "kube-api-access-2h9lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.759271 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-kube-api-access-gjbkg" (OuterVolumeSpecName: "kube-api-access-gjbkg") pod "68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" (UID: "68b331a3-a0a4-4fc8-bcd0-7d180f2a86be"). InnerVolumeSpecName "kube-api-access-gjbkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.840650 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-catalog-content\") pod \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.840731 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-utilities\") pod \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.840778 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rbws\" (UniqueName: \"kubernetes.io/projected/2908d4c0-bbe0-436b-b806-ca29cf20afcf-kube-api-access-2rbws\") pod \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.840804 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-catalog-content\") pod \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\" (UID: \"2908d4c0-bbe0-436b-b806-ca29cf20afcf\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.840839 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-catalog-content\") pod \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.840898 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-utilities\") pod \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\" (UID: \"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.840930 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d59rc\" (UniqueName: \"kubernetes.io/projected/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-kube-api-access-d59rc\") pod \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.840960 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-catalog-content\") pod \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\" (UID: \"35d2ebf2-6bd4-4adf-872d-d241a2519b0d\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.841022 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dwfh\" (UniqueName: \"kubernetes.io/projected/a5e65840-1d44-4cb3-bd42-3bc0048f7274-kube-api-access-5dwfh\") pod \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\" (UID: \"a5e65840-1d44-4cb3-bd42-3bc0048f7274\") " Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.841308 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h9lt\" (UniqueName: \"kubernetes.io/projected/5b3f76e2-8cce-47f1-87e5-43be0d07238e-kube-api-access-2h9lt\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.841342 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.841360 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.841356 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-utilities" (OuterVolumeSpecName: "utilities") pod "35d2ebf2-6bd4-4adf-872d-d241a2519b0d" (UID: "35d2ebf2-6bd4-4adf-872d-d241a2519b0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.841377 4789 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.841396 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjbkg\" (UniqueName: \"kubernetes.io/projected/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-kube-api-access-gjbkg\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.841415 4789 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/5b3f76e2-8cce-47f1-87e5-43be0d07238e-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.841618 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-utilities" (OuterVolumeSpecName: "utilities") pod "68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" (UID: "68b331a3-a0a4-4fc8-bcd0-7d180f2a86be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.843658 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2908d4c0-bbe0-436b-b806-ca29cf20afcf-kube-api-access-2rbws" (OuterVolumeSpecName: "kube-api-access-2rbws") pod "2908d4c0-bbe0-436b-b806-ca29cf20afcf" (UID: "2908d4c0-bbe0-436b-b806-ca29cf20afcf"). InnerVolumeSpecName "kube-api-access-2rbws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.844110 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-kube-api-access-d59rc" (OuterVolumeSpecName: "kube-api-access-d59rc") pod "35d2ebf2-6bd4-4adf-872d-d241a2519b0d" (UID: "35d2ebf2-6bd4-4adf-872d-d241a2519b0d"). InnerVolumeSpecName "kube-api-access-d59rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.844226 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e65840-1d44-4cb3-bd42-3bc0048f7274-kube-api-access-5dwfh" (OuterVolumeSpecName: "kube-api-access-5dwfh") pod "a5e65840-1d44-4cb3-bd42-3bc0048f7274" (UID: "a5e65840-1d44-4cb3-bd42-3bc0048f7274"). InnerVolumeSpecName "kube-api-access-5dwfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.857141 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35d2ebf2-6bd4-4adf-872d-d241a2519b0d" (UID: "35d2ebf2-6bd4-4adf-872d-d241a2519b0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.890883 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5e65840-1d44-4cb3-bd42-3bc0048f7274" (UID: "a5e65840-1d44-4cb3-bd42-3bc0048f7274"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.924662 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2908d4c0-bbe0-436b-b806-ca29cf20afcf" (UID: "2908d4c0-bbe0-436b-b806-ca29cf20afcf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.933252 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" (UID: "68b331a3-a0a4-4fc8-bcd0-7d180f2a86be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.941955 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.942016 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d59rc\" (UniqueName: \"kubernetes.io/projected/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-kube-api-access-d59rc\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.942033 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.942045 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dwfh\" (UniqueName: \"kubernetes.io/projected/a5e65840-1d44-4cb3-bd42-3bc0048f7274-kube-api-access-5dwfh\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.942057 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e65840-1d44-4cb3-bd42-3bc0048f7274-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.942068 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35d2ebf2-6bd4-4adf-872d-d241a2519b0d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.942081 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rbws\" (UniqueName: \"kubernetes.io/projected/2908d4c0-bbe0-436b-b806-ca29cf20afcf-kube-api-access-2rbws\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.942094 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2908d4c0-bbe0-436b-b806-ca29cf20afcf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:38 crc kubenswrapper[4789]: I1008 14:05:38.942106 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.015814 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4lxzv"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.272284 4789 generic.go:334] "Generic (PLEG): container finished" podID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerID="15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7" exitCode=0 Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.272630 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vn2fk" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.273027 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2fk" event={"ID":"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be","Type":"ContainerDied","Data":"15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.273098 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2fk" event={"ID":"68b331a3-a0a4-4fc8-bcd0-7d180f2a86be","Type":"ContainerDied","Data":"a8a62359b1b7d54218ec3d87223ebd5622772ea47a4a2927ba711b10c158c2a1"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.273123 4789 scope.go:117] "RemoveContainer" containerID="15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.274630 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" event={"ID":"c59c8d39-3840-48cf-b510-4e07d9293fc9","Type":"ContainerStarted","Data":"910ae5ceaebdf6dd89f4c97ad1ee0cf757b94dd1e4ca4a72539e7fc76217c7d2"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.274676 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" event={"ID":"c59c8d39-3840-48cf-b510-4e07d9293fc9","Type":"ContainerStarted","Data":"125ae789f7899275dbdac6c9348334d5f2a06d449880099c84110621d19b5827"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.275788 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.277331 4789 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4lxzv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.277379 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" podUID="c59c8d39-3840-48cf-b510-4e07d9293fc9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.278443 4789 generic.go:334] "Generic (PLEG): container finished" podID="5b3f76e2-8cce-47f1-87e5-43be0d07238e" containerID="dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf" exitCode=0 Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.278508 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" event={"ID":"5b3f76e2-8cce-47f1-87e5-43be0d07238e","Type":"ContainerDied","Data":"dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.278530 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" event={"ID":"5b3f76e2-8cce-47f1-87e5-43be0d07238e","Type":"ContainerDied","Data":"d37c36b6c06c86b617663fcacc4e1da5d66a2a20d62bee87bac3d374351ec4a0"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.278584 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vzq6g" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.288787 4789 generic.go:334] "Generic (PLEG): container finished" podID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerID="04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6" exitCode=0 Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.288878 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72nk5" event={"ID":"a5e65840-1d44-4cb3-bd42-3bc0048f7274","Type":"ContainerDied","Data":"04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.288917 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72nk5" event={"ID":"a5e65840-1d44-4cb3-bd42-3bc0048f7274","Type":"ContainerDied","Data":"00806708096f1a713137a29d99b0c1ab338848db74806ad452e2d9554ea2c18a"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.289037 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72nk5" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.292517 4789 generic.go:334] "Generic (PLEG): container finished" podID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerID="48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6" exitCode=0 Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.292587 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z84nj" event={"ID":"35d2ebf2-6bd4-4adf-872d-d241a2519b0d","Type":"ContainerDied","Data":"48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.292616 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z84nj" event={"ID":"35d2ebf2-6bd4-4adf-872d-d241a2519b0d","Type":"ContainerDied","Data":"b47b8039f0281bb7fe525583194bb9e3adb28b4abc2d189078f4e1cb3d7c570c"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.292690 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z84nj" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.299748 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" podStartSLOduration=1.299729546 podStartE2EDuration="1.299729546s" podCreationTimestamp="2025-10-08 14:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:05:39.298447811 +0000 UTC m=+279.205195323" watchObservedRunningTime="2025-10-08 14:05:39.299729546 +0000 UTC m=+279.206477058" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.304344 4789 generic.go:334] "Generic (PLEG): container finished" podID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerID="7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba" exitCode=0 Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.304389 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rswj8" event={"ID":"2908d4c0-bbe0-436b-b806-ca29cf20afcf","Type":"ContainerDied","Data":"7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.304416 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rswj8" event={"ID":"2908d4c0-bbe0-436b-b806-ca29cf20afcf","Type":"ContainerDied","Data":"cad15f30d5ee880ae26d8d1aca426e0b81871e3ea063f1d6ce3f4e9ef7a6b176"} Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.304423 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rswj8" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.345722 4789 scope.go:117] "RemoveContainer" containerID="5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.365781 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vzq6g"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.369628 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vzq6g"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.377227 4789 scope.go:117] "RemoveContainer" containerID="814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.387279 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72nk5"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.396480 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-72nk5"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.399566 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vn2fk"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.401656 4789 scope.go:117] "RemoveContainer" containerID="15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.404757 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7\": container with ID starting with 15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7 not found: ID does not exist" containerID="15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.404797 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7"} err="failed to get container status \"15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7\": rpc error: code = NotFound desc = could not find container \"15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7\": container with ID starting with 15f80c0dca998da407d2468af5a0d6f448449dd518c209c3368e64b48141b1e7 not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.404827 4789 scope.go:117] "RemoveContainer" containerID="5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.405297 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585\": container with ID starting with 5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585 not found: ID does not exist" containerID="5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.405324 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585"} err="failed to get container status \"5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585\": rpc error: code = NotFound desc = could not find container \"5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585\": container with ID starting with 5214e75cb10241b54ec4ca79e57084b0602340ff24fc7cc4c00fe12a5e592585 not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.405341 4789 scope.go:117] "RemoveContainer" containerID="814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.405603 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d\": container with ID starting with 814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d not found: ID does not exist" containerID="814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.405629 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d"} err="failed to get container status \"814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d\": rpc error: code = NotFound desc = could not find container \"814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d\": container with ID starting with 814c3a9d579d1af0518100c5f9926d683ecca01317918b441715c0056e72f94d not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.405645 4789 scope.go:117] "RemoveContainer" containerID="dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.408254 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vn2fk"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.423160 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z84nj"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.423275 4789 scope.go:117] "RemoveContainer" containerID="dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.423737 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf\": container with ID starting with dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf not found: ID does not exist" containerID="dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.423802 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf"} err="failed to get container status \"dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf\": rpc error: code = NotFound desc = could not find container \"dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf\": container with ID starting with dfb0fd1e59db22d9215c911aef3775238981a8cdfc69048bc5a26a6d685b71cf not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.423847 4789 scope.go:117] "RemoveContainer" containerID="04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.426235 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z84nj"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.429297 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rswj8"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.431315 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rswj8"] Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.437899 4789 scope.go:117] "RemoveContainer" containerID="128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.452680 4789 scope.go:117] "RemoveContainer" containerID="006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.466192 4789 scope.go:117] "RemoveContainer" containerID="04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.466628 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6\": container with ID starting with 04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6 not found: ID does not exist" containerID="04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.466737 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6"} err="failed to get container status \"04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6\": rpc error: code = NotFound desc = could not find container \"04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6\": container with ID starting with 04af2fccdab3df8e5cc2b400387f6c2a9029c765ecbaed44e0d943e05fdd1ff6 not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.466773 4789 scope.go:117] "RemoveContainer" containerID="128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.467138 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3\": container with ID starting with 128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3 not found: ID does not exist" containerID="128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.467209 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3"} err="failed to get container status \"128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3\": rpc error: code = NotFound desc = could not find container \"128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3\": container with ID starting with 128106742c5142587c54c334c23a5f0fd911026c995e764ffa180b4a26ad3df3 not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.467251 4789 scope.go:117] "RemoveContainer" containerID="006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.467564 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6\": container with ID starting with 006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6 not found: ID does not exist" containerID="006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.467588 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6"} err="failed to get container status \"006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6\": rpc error: code = NotFound desc = could not find container \"006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6\": container with ID starting with 006f74172560e7ff703e7f530600f2ed24e9ebb3a26c6748aef07cbb84b17ab6 not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.467622 4789 scope.go:117] "RemoveContainer" containerID="48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.479755 4789 scope.go:117] "RemoveContainer" containerID="e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.517806 4789 scope.go:117] "RemoveContainer" containerID="d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.532112 4789 scope.go:117] "RemoveContainer" containerID="48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.532509 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6\": container with ID starting with 48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6 not found: ID does not exist" containerID="48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.532562 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6"} err="failed to get container status \"48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6\": rpc error: code = NotFound desc = could not find container \"48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6\": container with ID starting with 48a3e5065ff9db1723199286d8750088b99cdfb71920fbd1a0046ec7898847d6 not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.532603 4789 scope.go:117] "RemoveContainer" containerID="e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.533043 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337\": container with ID starting with e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337 not found: ID does not exist" containerID="e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.533097 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337"} err="failed to get container status \"e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337\": rpc error: code = NotFound desc = could not find container \"e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337\": container with ID starting with e8c1e3aa05727c947d7632dede1e680ceb35cabce96845f066298aff93076337 not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.533129 4789 scope.go:117] "RemoveContainer" containerID="d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.533379 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288\": container with ID starting with d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288 not found: ID does not exist" containerID="d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.533416 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288"} err="failed to get container status \"d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288\": rpc error: code = NotFound desc = could not find container \"d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288\": container with ID starting with d2212c6dcc0270355ffe1a4c8367339a7fa159e7bd768c75f725186592e9f288 not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.533431 4789 scope.go:117] "RemoveContainer" containerID="7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.550837 4789 scope.go:117] "RemoveContainer" containerID="7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.565317 4789 scope.go:117] "RemoveContainer" containerID="00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.583562 4789 scope.go:117] "RemoveContainer" containerID="7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.584109 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba\": container with ID starting with 7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba not found: ID does not exist" containerID="7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.584181 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba"} err="failed to get container status \"7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba\": rpc error: code = NotFound desc = could not find container \"7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba\": container with ID starting with 7deab406d96364fdd6a8831d7f2b2198675cd4d58f5944cb24f436cb121d3eba not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.584215 4789 scope.go:117] "RemoveContainer" containerID="7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.584844 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49\": container with ID starting with 7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49 not found: ID does not exist" containerID="7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.584874 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49"} err="failed to get container status \"7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49\": rpc error: code = NotFound desc = could not find container \"7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49\": container with ID starting with 7b196f2ca34558e42899dc37a6933362ecb520a3bb4d6519d9e854834ce5db49 not found: ID does not exist" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.584890 4789 scope.go:117] "RemoveContainer" containerID="00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185" Oct 08 14:05:39 crc kubenswrapper[4789]: E1008 14:05:39.585590 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185\": container with ID starting with 00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185 not found: ID does not exist" containerID="00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185" Oct 08 14:05:39 crc kubenswrapper[4789]: I1008 14:05:39.585636 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185"} err="failed to get container status \"00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185\": rpc error: code = NotFound desc = could not find container \"00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185\": container with ID starting with 00888ad9241d5ba92d88f32128cff023b12b7f388ce5f9c8f5accd47c0d2c185 not found: ID does not exist" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.147937 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gxpd9"] Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148130 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148142 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148151 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerName="extract-content" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148157 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerName="extract-content" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148163 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148169 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148180 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerName="extract-content" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148185 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerName="extract-content" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148192 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerName="extract-content" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148197 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerName="extract-content" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148205 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148212 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148224 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerName="extract-utilities" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148229 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerName="extract-utilities" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148238 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148244 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148253 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerName="extract-content" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148259 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerName="extract-content" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148267 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerName="extract-utilities" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148272 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerName="extract-utilities" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148279 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3f76e2-8cce-47f1-87e5-43be0d07238e" containerName="marketplace-operator" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148285 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3f76e2-8cce-47f1-87e5-43be0d07238e" containerName="marketplace-operator" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148297 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerName="extract-utilities" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148303 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerName="extract-utilities" Oct 08 14:05:40 crc kubenswrapper[4789]: E1008 14:05:40.148309 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerName="extract-utilities" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148316 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerName="extract-utilities" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148405 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b3f76e2-8cce-47f1-87e5-43be0d07238e" containerName="marketplace-operator" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148417 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148427 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148437 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.148444 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" containerName="registry-server" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.149149 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.150870 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.155095 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d530161d-7c51-48bb-8c08-0f48a5842e3e-utilities\") pod \"redhat-marketplace-gxpd9\" (UID: \"d530161d-7c51-48bb-8c08-0f48a5842e3e\") " pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.155176 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxh2s\" (UniqueName: \"kubernetes.io/projected/d530161d-7c51-48bb-8c08-0f48a5842e3e-kube-api-access-sxh2s\") pod \"redhat-marketplace-gxpd9\" (UID: \"d530161d-7c51-48bb-8c08-0f48a5842e3e\") " pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.155209 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d530161d-7c51-48bb-8c08-0f48a5842e3e-catalog-content\") pod \"redhat-marketplace-gxpd9\" (UID: \"d530161d-7c51-48bb-8c08-0f48a5842e3e\") " pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.161199 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxpd9"] Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.256049 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d530161d-7c51-48bb-8c08-0f48a5842e3e-catalog-content\") pod \"redhat-marketplace-gxpd9\" (UID: \"d530161d-7c51-48bb-8c08-0f48a5842e3e\") " pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.256127 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d530161d-7c51-48bb-8c08-0f48a5842e3e-utilities\") pod \"redhat-marketplace-gxpd9\" (UID: \"d530161d-7c51-48bb-8c08-0f48a5842e3e\") " pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.256358 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxh2s\" (UniqueName: \"kubernetes.io/projected/d530161d-7c51-48bb-8c08-0f48a5842e3e-kube-api-access-sxh2s\") pod \"redhat-marketplace-gxpd9\" (UID: \"d530161d-7c51-48bb-8c08-0f48a5842e3e\") " pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.256637 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d530161d-7c51-48bb-8c08-0f48a5842e3e-utilities\") pod \"redhat-marketplace-gxpd9\" (UID: \"d530161d-7c51-48bb-8c08-0f48a5842e3e\") " pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.256773 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d530161d-7c51-48bb-8c08-0f48a5842e3e-catalog-content\") pod \"redhat-marketplace-gxpd9\" (UID: \"d530161d-7c51-48bb-8c08-0f48a5842e3e\") " pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.275679 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxh2s\" (UniqueName: \"kubernetes.io/projected/d530161d-7c51-48bb-8c08-0f48a5842e3e-kube-api-access-sxh2s\") pod \"redhat-marketplace-gxpd9\" (UID: \"d530161d-7c51-48bb-8c08-0f48a5842e3e\") " pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.319889 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4lxzv" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.464851 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.659850 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxpd9"] Oct 08 14:05:40 crc kubenswrapper[4789]: W1008 14:05:40.665618 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd530161d_7c51_48bb_8c08_0f48a5842e3e.slice/crio-9795e91c66c590eb78b08654d5769c336e7fe6596db029197ff061df02b4ec23 WatchSource:0}: Error finding container 9795e91c66c590eb78b08654d5769c336e7fe6596db029197ff061df02b4ec23: Status 404 returned error can't find the container with id 9795e91c66c590eb78b08654d5769c336e7fe6596db029197ff061df02b4ec23 Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.749062 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2908d4c0-bbe0-436b-b806-ca29cf20afcf" path="/var/lib/kubelet/pods/2908d4c0-bbe0-436b-b806-ca29cf20afcf/volumes" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.752980 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d2ebf2-6bd4-4adf-872d-d241a2519b0d" path="/var/lib/kubelet/pods/35d2ebf2-6bd4-4adf-872d-d241a2519b0d/volumes" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.754458 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b3f76e2-8cce-47f1-87e5-43be0d07238e" path="/var/lib/kubelet/pods/5b3f76e2-8cce-47f1-87e5-43be0d07238e/volumes" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.756086 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b331a3-a0a4-4fc8-bcd0-7d180f2a86be" path="/var/lib/kubelet/pods/68b331a3-a0a4-4fc8-bcd0-7d180f2a86be/volumes" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.756792 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5e65840-1d44-4cb3-bd42-3bc0048f7274" path="/var/lib/kubelet/pods/a5e65840-1d44-4cb3-bd42-3bc0048f7274/volumes" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.757458 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dlsnl"] Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.758630 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.761056 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.761087 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dlsnl"] Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.761597 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-utilities\") pod \"certified-operators-dlsnl\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.761687 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-catalog-content\") pod \"certified-operators-dlsnl\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.761729 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n44wz\" (UniqueName: \"kubernetes.io/projected/1695b8b4-119e-4626-8e40-5ecd20d08259-kube-api-access-n44wz\") pod \"certified-operators-dlsnl\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.862144 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-catalog-content\") pod \"certified-operators-dlsnl\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.862185 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n44wz\" (UniqueName: \"kubernetes.io/projected/1695b8b4-119e-4626-8e40-5ecd20d08259-kube-api-access-n44wz\") pod \"certified-operators-dlsnl\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.862224 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-utilities\") pod \"certified-operators-dlsnl\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.862612 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-catalog-content\") pod \"certified-operators-dlsnl\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.862631 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-utilities\") pod \"certified-operators-dlsnl\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:40 crc kubenswrapper[4789]: I1008 14:05:40.891312 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n44wz\" (UniqueName: \"kubernetes.io/projected/1695b8b4-119e-4626-8e40-5ecd20d08259-kube-api-access-n44wz\") pod \"certified-operators-dlsnl\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:41 crc kubenswrapper[4789]: I1008 14:05:41.074551 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:41 crc kubenswrapper[4789]: I1008 14:05:41.251371 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dlsnl"] Oct 08 14:05:41 crc kubenswrapper[4789]: I1008 14:05:41.322578 4789 generic.go:334] "Generic (PLEG): container finished" podID="d530161d-7c51-48bb-8c08-0f48a5842e3e" containerID="99807b39462d9963a7c0ec359678313aa89600e37353aba61c7c85464857d401" exitCode=0 Oct 08 14:05:41 crc kubenswrapper[4789]: I1008 14:05:41.322637 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxpd9" event={"ID":"d530161d-7c51-48bb-8c08-0f48a5842e3e","Type":"ContainerDied","Data":"99807b39462d9963a7c0ec359678313aa89600e37353aba61c7c85464857d401"} Oct 08 14:05:41 crc kubenswrapper[4789]: I1008 14:05:41.322666 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxpd9" event={"ID":"d530161d-7c51-48bb-8c08-0f48a5842e3e","Type":"ContainerStarted","Data":"9795e91c66c590eb78b08654d5769c336e7fe6596db029197ff061df02b4ec23"} Oct 08 14:05:41 crc kubenswrapper[4789]: I1008 14:05:41.323859 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlsnl" event={"ID":"1695b8b4-119e-4626-8e40-5ecd20d08259","Type":"ContainerStarted","Data":"8d85482fd83b89e47357f279edae0a9b77cf10abd681b2ec328b588e30fa8d40"} Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.332268 4789 generic.go:334] "Generic (PLEG): container finished" podID="d530161d-7c51-48bb-8c08-0f48a5842e3e" containerID="7d1e13d484e559db92bf4ee6f945787d8b5af0189308cda913fc5ac768637aba" exitCode=0 Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.332344 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxpd9" event={"ID":"d530161d-7c51-48bb-8c08-0f48a5842e3e","Type":"ContainerDied","Data":"7d1e13d484e559db92bf4ee6f945787d8b5af0189308cda913fc5ac768637aba"} Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.336555 4789 generic.go:334] "Generic (PLEG): container finished" podID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerID="58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a" exitCode=0 Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.336788 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlsnl" event={"ID":"1695b8b4-119e-4626-8e40-5ecd20d08259","Type":"ContainerDied","Data":"58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a"} Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.549828 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pqnwd"] Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.550983 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.553475 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.562133 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqnwd"] Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.587728 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51eae247-9044-49e7-a270-6a943c12a3ac-utilities\") pod \"redhat-operators-pqnwd\" (UID: \"51eae247-9044-49e7-a270-6a943c12a3ac\") " pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.587816 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89ttl\" (UniqueName: \"kubernetes.io/projected/51eae247-9044-49e7-a270-6a943c12a3ac-kube-api-access-89ttl\") pod \"redhat-operators-pqnwd\" (UID: \"51eae247-9044-49e7-a270-6a943c12a3ac\") " pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.587868 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51eae247-9044-49e7-a270-6a943c12a3ac-catalog-content\") pod \"redhat-operators-pqnwd\" (UID: \"51eae247-9044-49e7-a270-6a943c12a3ac\") " pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.688683 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51eae247-9044-49e7-a270-6a943c12a3ac-catalog-content\") pod \"redhat-operators-pqnwd\" (UID: \"51eae247-9044-49e7-a270-6a943c12a3ac\") " pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.688773 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51eae247-9044-49e7-a270-6a943c12a3ac-utilities\") pod \"redhat-operators-pqnwd\" (UID: \"51eae247-9044-49e7-a270-6a943c12a3ac\") " pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.688833 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89ttl\" (UniqueName: \"kubernetes.io/projected/51eae247-9044-49e7-a270-6a943c12a3ac-kube-api-access-89ttl\") pod \"redhat-operators-pqnwd\" (UID: \"51eae247-9044-49e7-a270-6a943c12a3ac\") " pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.689214 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51eae247-9044-49e7-a270-6a943c12a3ac-catalog-content\") pod \"redhat-operators-pqnwd\" (UID: \"51eae247-9044-49e7-a270-6a943c12a3ac\") " pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.689250 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51eae247-9044-49e7-a270-6a943c12a3ac-utilities\") pod \"redhat-operators-pqnwd\" (UID: \"51eae247-9044-49e7-a270-6a943c12a3ac\") " pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.709215 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89ttl\" (UniqueName: \"kubernetes.io/projected/51eae247-9044-49e7-a270-6a943c12a3ac-kube-api-access-89ttl\") pod \"redhat-operators-pqnwd\" (UID: \"51eae247-9044-49e7-a270-6a943c12a3ac\") " pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:42 crc kubenswrapper[4789]: I1008 14:05:42.867054 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.145416 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-flhdr"] Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.146858 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.148792 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.157490 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-flhdr"] Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.194146 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-utilities\") pod \"community-operators-flhdr\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.194207 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-catalog-content\") pod \"community-operators-flhdr\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.194276 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g55b\" (UniqueName: \"kubernetes.io/projected/673f91dd-cda6-4e43-813f-644ac69de215-kube-api-access-7g55b\") pod \"community-operators-flhdr\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.295071 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g55b\" (UniqueName: \"kubernetes.io/projected/673f91dd-cda6-4e43-813f-644ac69de215-kube-api-access-7g55b\") pod \"community-operators-flhdr\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.295383 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-utilities\") pod \"community-operators-flhdr\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.295562 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-catalog-content\") pod \"community-operators-flhdr\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.295799 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-utilities\") pod \"community-operators-flhdr\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.296049 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-catalog-content\") pod \"community-operators-flhdr\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.300603 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqnwd"] Oct 08 14:05:43 crc kubenswrapper[4789]: W1008 14:05:43.303816 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51eae247_9044_49e7_a270_6a943c12a3ac.slice/crio-a384ba5f0a7156739a8597dde0425a8fda8da65ef9ab2927a749438cd5fb8a77 WatchSource:0}: Error finding container a384ba5f0a7156739a8597dde0425a8fda8da65ef9ab2927a749438cd5fb8a77: Status 404 returned error can't find the container with id a384ba5f0a7156739a8597dde0425a8fda8da65ef9ab2927a749438cd5fb8a77 Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.319937 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g55b\" (UniqueName: \"kubernetes.io/projected/673f91dd-cda6-4e43-813f-644ac69de215-kube-api-access-7g55b\") pod \"community-operators-flhdr\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.343999 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxpd9" event={"ID":"d530161d-7c51-48bb-8c08-0f48a5842e3e","Type":"ContainerStarted","Data":"9ad439e99a6ccef1dfd2a763c28d0778e14f6de26f3d45a6320d31b9d88cae49"} Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.348384 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlsnl" event={"ID":"1695b8b4-119e-4626-8e40-5ecd20d08259","Type":"ContainerStarted","Data":"b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560"} Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.349463 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqnwd" event={"ID":"51eae247-9044-49e7-a270-6a943c12a3ac","Type":"ContainerStarted","Data":"a384ba5f0a7156739a8597dde0425a8fda8da65ef9ab2927a749438cd5fb8a77"} Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.363817 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gxpd9" podStartSLOduration=1.8884170340000002 podStartE2EDuration="3.363797818s" podCreationTimestamp="2025-10-08 14:05:40 +0000 UTC" firstStartedPulling="2025-10-08 14:05:41.323955549 +0000 UTC m=+281.230703041" lastFinishedPulling="2025-10-08 14:05:42.799336333 +0000 UTC m=+282.706083825" observedRunningTime="2025-10-08 14:05:43.362481182 +0000 UTC m=+283.269228684" watchObservedRunningTime="2025-10-08 14:05:43.363797818 +0000 UTC m=+283.270545310" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.469227 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:43 crc kubenswrapper[4789]: I1008 14:05:43.662555 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-flhdr"] Oct 08 14:05:44 crc kubenswrapper[4789]: I1008 14:05:44.362350 4789 generic.go:334] "Generic (PLEG): container finished" podID="51eae247-9044-49e7-a270-6a943c12a3ac" containerID="2c2d5ff29ab06a1bc55996b2130af947e459b82651c4db7acb9b929d98a344b1" exitCode=0 Oct 08 14:05:44 crc kubenswrapper[4789]: I1008 14:05:44.362398 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqnwd" event={"ID":"51eae247-9044-49e7-a270-6a943c12a3ac","Type":"ContainerDied","Data":"2c2d5ff29ab06a1bc55996b2130af947e459b82651c4db7acb9b929d98a344b1"} Oct 08 14:05:44 crc kubenswrapper[4789]: I1008 14:05:44.365409 4789 generic.go:334] "Generic (PLEG): container finished" podID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerID="b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560" exitCode=0 Oct 08 14:05:44 crc kubenswrapper[4789]: I1008 14:05:44.365458 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlsnl" event={"ID":"1695b8b4-119e-4626-8e40-5ecd20d08259","Type":"ContainerDied","Data":"b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560"} Oct 08 14:05:44 crc kubenswrapper[4789]: I1008 14:05:44.365479 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlsnl" event={"ID":"1695b8b4-119e-4626-8e40-5ecd20d08259","Type":"ContainerStarted","Data":"ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89"} Oct 08 14:05:44 crc kubenswrapper[4789]: I1008 14:05:44.369343 4789 generic.go:334] "Generic (PLEG): container finished" podID="673f91dd-cda6-4e43-813f-644ac69de215" containerID="bf6341b95af736f0493a4a252079c4d39846a599165c41996c7b004273c39b14" exitCode=0 Oct 08 14:05:44 crc kubenswrapper[4789]: I1008 14:05:44.369382 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flhdr" event={"ID":"673f91dd-cda6-4e43-813f-644ac69de215","Type":"ContainerDied","Data":"bf6341b95af736f0493a4a252079c4d39846a599165c41996c7b004273c39b14"} Oct 08 14:05:44 crc kubenswrapper[4789]: I1008 14:05:44.369419 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flhdr" event={"ID":"673f91dd-cda6-4e43-813f-644ac69de215","Type":"ContainerStarted","Data":"6ce5358441185d3f7eb02b31e193d47c1d05346aed716cf1982f79eedcf9b9eb"} Oct 08 14:05:44 crc kubenswrapper[4789]: I1008 14:05:44.419431 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dlsnl" podStartSLOduration=2.982893766 podStartE2EDuration="4.41941705s" podCreationTimestamp="2025-10-08 14:05:40 +0000 UTC" firstStartedPulling="2025-10-08 14:05:42.338679107 +0000 UTC m=+282.245426599" lastFinishedPulling="2025-10-08 14:05:43.775202401 +0000 UTC m=+283.681949883" observedRunningTime="2025-10-08 14:05:44.417237501 +0000 UTC m=+284.323984993" watchObservedRunningTime="2025-10-08 14:05:44.41941705 +0000 UTC m=+284.326164542" Oct 08 14:05:45 crc kubenswrapper[4789]: I1008 14:05:45.375512 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqnwd" event={"ID":"51eae247-9044-49e7-a270-6a943c12a3ac","Type":"ContainerStarted","Data":"4f82efc29478371e949e98ff8fb384475be91281cb2878c774f4509987fe8924"} Oct 08 14:05:46 crc kubenswrapper[4789]: I1008 14:05:46.382701 4789 generic.go:334] "Generic (PLEG): container finished" podID="51eae247-9044-49e7-a270-6a943c12a3ac" containerID="4f82efc29478371e949e98ff8fb384475be91281cb2878c774f4509987fe8924" exitCode=0 Oct 08 14:05:46 crc kubenswrapper[4789]: I1008 14:05:46.382805 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqnwd" event={"ID":"51eae247-9044-49e7-a270-6a943c12a3ac","Type":"ContainerDied","Data":"4f82efc29478371e949e98ff8fb384475be91281cb2878c774f4509987fe8924"} Oct 08 14:05:47 crc kubenswrapper[4789]: I1008 14:05:47.389577 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqnwd" event={"ID":"51eae247-9044-49e7-a270-6a943c12a3ac","Type":"ContainerStarted","Data":"f4b1e00a4654c01e88ddd3fbfa88c02b3d4cce9c69e4d76f267b3b959625d85d"} Oct 08 14:05:47 crc kubenswrapper[4789]: I1008 14:05:47.392100 4789 generic.go:334] "Generic (PLEG): container finished" podID="673f91dd-cda6-4e43-813f-644ac69de215" containerID="b9ec20d356e069830cb0a43fcf1293d17a0d80d4490088f396ceefbe446ef6af" exitCode=0 Oct 08 14:05:47 crc kubenswrapper[4789]: I1008 14:05:47.392127 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flhdr" event={"ID":"673f91dd-cda6-4e43-813f-644ac69de215","Type":"ContainerDied","Data":"b9ec20d356e069830cb0a43fcf1293d17a0d80d4490088f396ceefbe446ef6af"} Oct 08 14:05:47 crc kubenswrapper[4789]: I1008 14:05:47.409468 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pqnwd" podStartSLOduration=2.863270588 podStartE2EDuration="5.409445328s" podCreationTimestamp="2025-10-08 14:05:42 +0000 UTC" firstStartedPulling="2025-10-08 14:05:44.364486403 +0000 UTC m=+284.271233915" lastFinishedPulling="2025-10-08 14:05:46.910661163 +0000 UTC m=+286.817408655" observedRunningTime="2025-10-08 14:05:47.406541729 +0000 UTC m=+287.313289221" watchObservedRunningTime="2025-10-08 14:05:47.409445328 +0000 UTC m=+287.316192820" Oct 08 14:05:49 crc kubenswrapper[4789]: I1008 14:05:49.404250 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flhdr" event={"ID":"673f91dd-cda6-4e43-813f-644ac69de215","Type":"ContainerStarted","Data":"ee44700090d88c7d7084272dd9d8e650b609188d60e5dcc2fb28fc71cf86aef8"} Oct 08 14:05:49 crc kubenswrapper[4789]: I1008 14:05:49.433096 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-flhdr" podStartSLOduration=2.519690061 podStartE2EDuration="6.433079416s" podCreationTimestamp="2025-10-08 14:05:43 +0000 UTC" firstStartedPulling="2025-10-08 14:05:44.371364891 +0000 UTC m=+284.278112393" lastFinishedPulling="2025-10-08 14:05:48.284754246 +0000 UTC m=+288.191501748" observedRunningTime="2025-10-08 14:05:49.4317453 +0000 UTC m=+289.338492792" watchObservedRunningTime="2025-10-08 14:05:49.433079416 +0000 UTC m=+289.339826898" Oct 08 14:05:50 crc kubenswrapper[4789]: I1008 14:05:50.466267 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:50 crc kubenswrapper[4789]: I1008 14:05:50.466334 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:50 crc kubenswrapper[4789]: I1008 14:05:50.501794 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:51 crc kubenswrapper[4789]: I1008 14:05:51.075571 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:51 crc kubenswrapper[4789]: I1008 14:05:51.075624 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:51 crc kubenswrapper[4789]: I1008 14:05:51.133699 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:51 crc kubenswrapper[4789]: I1008 14:05:51.451382 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gxpd9" Oct 08 14:05:51 crc kubenswrapper[4789]: I1008 14:05:51.457593 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 14:05:52 crc kubenswrapper[4789]: I1008 14:05:52.867453 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:52 crc kubenswrapper[4789]: I1008 14:05:52.869732 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:52 crc kubenswrapper[4789]: I1008 14:05:52.925480 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:53 crc kubenswrapper[4789]: I1008 14:05:53.470764 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:53 crc kubenswrapper[4789]: I1008 14:05:53.471070 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:53 crc kubenswrapper[4789]: I1008 14:05:53.481616 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pqnwd" Oct 08 14:05:53 crc kubenswrapper[4789]: I1008 14:05:53.534153 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:05:54 crc kubenswrapper[4789]: I1008 14:05:54.489161 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-flhdr" Oct 08 14:06:56 crc kubenswrapper[4789]: I1008 14:06:56.432504 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:06:56 crc kubenswrapper[4789]: I1008 14:06:56.433196 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:07:26 crc kubenswrapper[4789]: I1008 14:07:26.432564 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:07:26 crc kubenswrapper[4789]: I1008 14:07:26.433239 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:07:56 crc kubenswrapper[4789]: I1008 14:07:56.432865 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:07:56 crc kubenswrapper[4789]: I1008 14:07:56.434096 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:07:56 crc kubenswrapper[4789]: I1008 14:07:56.434188 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:07:56 crc kubenswrapper[4789]: I1008 14:07:56.435410 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c481a44e4e2f4fdaae1b3ca23f477eaf282c238b6ce960c39dd22247b81d0297"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:07:56 crc kubenswrapper[4789]: I1008 14:07:56.435536 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://c481a44e4e2f4fdaae1b3ca23f477eaf282c238b6ce960c39dd22247b81d0297" gracePeriod=600 Oct 08 14:07:57 crc kubenswrapper[4789]: I1008 14:07:57.180287 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="c481a44e4e2f4fdaae1b3ca23f477eaf282c238b6ce960c39dd22247b81d0297" exitCode=0 Oct 08 14:07:57 crc kubenswrapper[4789]: I1008 14:07:57.180383 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"c481a44e4e2f4fdaae1b3ca23f477eaf282c238b6ce960c39dd22247b81d0297"} Oct 08 14:07:57 crc kubenswrapper[4789]: I1008 14:07:57.180796 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"dc34d79cea1ff807429be5448bf01d23f153ad3a6633f32da4bad1ac75061617"} Oct 08 14:07:57 crc kubenswrapper[4789]: I1008 14:07:57.180828 4789 scope.go:117] "RemoveContainer" containerID="e956e011adf58d071b3f26b2c0b916c21e9dfe6c91c1eba0400285062484e68f" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.297679 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tnr7v"] Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.299143 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.320149 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tnr7v"] Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.428786 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5r62\" (UniqueName: \"kubernetes.io/projected/beb65903-8585-4a20-94bb-cc6b8ca248b3-kube-api-access-n5r62\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.428831 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beb65903-8585-4a20-94bb-cc6b8ca248b3-trusted-ca\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.428865 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/beb65903-8585-4a20-94bb-cc6b8ca248b3-registry-certificates\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.428884 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/beb65903-8585-4a20-94bb-cc6b8ca248b3-registry-tls\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.428899 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/beb65903-8585-4a20-94bb-cc6b8ca248b3-bound-sa-token\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.429034 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/beb65903-8585-4a20-94bb-cc6b8ca248b3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.429149 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/beb65903-8585-4a20-94bb-cc6b8ca248b3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.429184 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.447390 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.530809 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beb65903-8585-4a20-94bb-cc6b8ca248b3-trusted-ca\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.530878 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/beb65903-8585-4a20-94bb-cc6b8ca248b3-registry-certificates\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.530903 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/beb65903-8585-4a20-94bb-cc6b8ca248b3-registry-tls\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.530921 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/beb65903-8585-4a20-94bb-cc6b8ca248b3-bound-sa-token\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.530972 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/beb65903-8585-4a20-94bb-cc6b8ca248b3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.531370 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/beb65903-8585-4a20-94bb-cc6b8ca248b3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.531738 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/beb65903-8585-4a20-94bb-cc6b8ca248b3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.531811 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5r62\" (UniqueName: \"kubernetes.io/projected/beb65903-8585-4a20-94bb-cc6b8ca248b3-kube-api-access-n5r62\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.532415 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/beb65903-8585-4a20-94bb-cc6b8ca248b3-registry-certificates\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.532861 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/beb65903-8585-4a20-94bb-cc6b8ca248b3-trusted-ca\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.536366 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/beb65903-8585-4a20-94bb-cc6b8ca248b3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.536497 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/beb65903-8585-4a20-94bb-cc6b8ca248b3-registry-tls\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.554693 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/beb65903-8585-4a20-94bb-cc6b8ca248b3-bound-sa-token\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.555297 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5r62\" (UniqueName: \"kubernetes.io/projected/beb65903-8585-4a20-94bb-cc6b8ca248b3-kube-api-access-n5r62\") pod \"image-registry-66df7c8f76-tnr7v\" (UID: \"beb65903-8585-4a20-94bb-cc6b8ca248b3\") " pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:20 crc kubenswrapper[4789]: I1008 14:09:20.614494 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:21 crc kubenswrapper[4789]: I1008 14:09:21.026654 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tnr7v"] Oct 08 14:09:21 crc kubenswrapper[4789]: I1008 14:09:21.658493 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" event={"ID":"beb65903-8585-4a20-94bb-cc6b8ca248b3","Type":"ContainerStarted","Data":"39666ded31e336a41ac854ed5a3065643e29350af8043c24afb7fc71320490cd"} Oct 08 14:09:21 crc kubenswrapper[4789]: I1008 14:09:21.658551 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" event={"ID":"beb65903-8585-4a20-94bb-cc6b8ca248b3","Type":"ContainerStarted","Data":"d51490ca13c68ce3aa5d15c1a2c375921a54b2deb896f860bb577bf99dc26aa4"} Oct 08 14:09:21 crc kubenswrapper[4789]: I1008 14:09:21.658664 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:40 crc kubenswrapper[4789]: I1008 14:09:40.622313 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" Oct 08 14:09:40 crc kubenswrapper[4789]: I1008 14:09:40.649178 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-tnr7v" podStartSLOduration=20.649148285 podStartE2EDuration="20.649148285s" podCreationTimestamp="2025-10-08 14:09:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:09:21.683643814 +0000 UTC m=+501.590391316" watchObservedRunningTime="2025-10-08 14:09:40.649148285 +0000 UTC m=+520.555895817" Oct 08 14:09:40 crc kubenswrapper[4789]: I1008 14:09:40.755910 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mc249"] Oct 08 14:09:56 crc kubenswrapper[4789]: I1008 14:09:56.432558 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:09:56 crc kubenswrapper[4789]: I1008 14:09:56.433253 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:10:05 crc kubenswrapper[4789]: I1008 14:10:05.790430 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" podUID="d124aeeb-8b80-4c21-b9c5-617be3ed17f6" containerName="registry" containerID="cri-o://b5f4ced7f877ef91a2daca300c3de8f3d2e9a1495e1800c87524af7796a9be59" gracePeriod=30 Oct 08 14:10:05 crc kubenswrapper[4789]: I1008 14:10:05.916603 4789 generic.go:334] "Generic (PLEG): container finished" podID="d124aeeb-8b80-4c21-b9c5-617be3ed17f6" containerID="b5f4ced7f877ef91a2daca300c3de8f3d2e9a1495e1800c87524af7796a9be59" exitCode=0 Oct 08 14:10:05 crc kubenswrapper[4789]: I1008 14:10:05.916653 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" event={"ID":"d124aeeb-8b80-4c21-b9c5-617be3ed17f6","Type":"ContainerDied","Data":"b5f4ced7f877ef91a2daca300c3de8f3d2e9a1495e1800c87524af7796a9be59"} Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.144218 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.187910 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-certificates\") pod \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.188008 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-tls\") pod \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.188057 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-installation-pull-secrets\") pod \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.188157 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-ca-trust-extracted\") pod \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.188314 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.188350 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-bound-sa-token\") pod \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.188415 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q2mw\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-kube-api-access-8q2mw\") pod \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.188459 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-trusted-ca\") pod \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\" (UID: \"d124aeeb-8b80-4c21-b9c5-617be3ed17f6\") " Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.189127 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d124aeeb-8b80-4c21-b9c5-617be3ed17f6" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.189911 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d124aeeb-8b80-4c21-b9c5-617be3ed17f6" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.195195 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d124aeeb-8b80-4c21-b9c5-617be3ed17f6" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.195349 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d124aeeb-8b80-4c21-b9c5-617be3ed17f6" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.195945 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d124aeeb-8b80-4c21-b9c5-617be3ed17f6" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.196286 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-kube-api-access-8q2mw" (OuterVolumeSpecName: "kube-api-access-8q2mw") pod "d124aeeb-8b80-4c21-b9c5-617be3ed17f6" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6"). InnerVolumeSpecName "kube-api-access-8q2mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.202669 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d124aeeb-8b80-4c21-b9c5-617be3ed17f6" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.210818 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d124aeeb-8b80-4c21-b9c5-617be3ed17f6" (UID: "d124aeeb-8b80-4c21-b9c5-617be3ed17f6"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.289716 4789 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.289756 4789 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.289769 4789 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.289780 4789 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.289791 4789 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.289802 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q2mw\" (UniqueName: \"kubernetes.io/projected/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-kube-api-access-8q2mw\") on node \"crc\" DevicePath \"\"" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.289812 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d124aeeb-8b80-4c21-b9c5-617be3ed17f6-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.922671 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" event={"ID":"d124aeeb-8b80-4c21-b9c5-617be3ed17f6","Type":"ContainerDied","Data":"f77aa9982285ad5a0dda5c6b4d7aee213deb7f4e00eda7baa3e35360245e5de5"} Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.922723 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mc249" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.922749 4789 scope.go:117] "RemoveContainer" containerID="b5f4ced7f877ef91a2daca300c3de8f3d2e9a1495e1800c87524af7796a9be59" Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.937104 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mc249"] Oct 08 14:10:06 crc kubenswrapper[4789]: I1008 14:10:06.942743 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mc249"] Oct 08 14:10:08 crc kubenswrapper[4789]: I1008 14:10:08.741691 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d124aeeb-8b80-4c21-b9c5-617be3ed17f6" path="/var/lib/kubelet/pods/d124aeeb-8b80-4c21-b9c5-617be3ed17f6/volumes" Oct 08 14:10:26 crc kubenswrapper[4789]: I1008 14:10:26.432697 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:10:26 crc kubenswrapper[4789]: I1008 14:10:26.433197 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:10:56 crc kubenswrapper[4789]: I1008 14:10:56.432938 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:10:56 crc kubenswrapper[4789]: I1008 14:10:56.433571 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:10:56 crc kubenswrapper[4789]: I1008 14:10:56.433618 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:10:56 crc kubenswrapper[4789]: I1008 14:10:56.434315 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc34d79cea1ff807429be5448bf01d23f153ad3a6633f32da4bad1ac75061617"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:10:56 crc kubenswrapper[4789]: I1008 14:10:56.434379 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://dc34d79cea1ff807429be5448bf01d23f153ad3a6633f32da4bad1ac75061617" gracePeriod=600 Oct 08 14:10:57 crc kubenswrapper[4789]: I1008 14:10:57.217063 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="dc34d79cea1ff807429be5448bf01d23f153ad3a6633f32da4bad1ac75061617" exitCode=0 Oct 08 14:10:57 crc kubenswrapper[4789]: I1008 14:10:57.217132 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"dc34d79cea1ff807429be5448bf01d23f153ad3a6633f32da4bad1ac75061617"} Oct 08 14:10:57 crc kubenswrapper[4789]: I1008 14:10:57.217370 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"8158c6bcd8d574426927fe18b85723afd7f1a9d1d3d0dd0ab1d58a6c4269820d"} Oct 08 14:10:57 crc kubenswrapper[4789]: I1008 14:10:57.217387 4789 scope.go:117] "RemoveContainer" containerID="c481a44e4e2f4fdaae1b3ca23f477eaf282c238b6ce960c39dd22247b81d0297" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.175328 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-5vfnw"] Oct 08 14:11:46 crc kubenswrapper[4789]: E1008 14:11:46.176153 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d124aeeb-8b80-4c21-b9c5-617be3ed17f6" containerName="registry" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.176170 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d124aeeb-8b80-4c21-b9c5-617be3ed17f6" containerName="registry" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.176311 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d124aeeb-8b80-4c21-b9c5-617be3ed17f6" containerName="registry" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.176757 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-5vfnw" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.179119 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.179253 4789 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-cjjnb" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.179307 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.186351 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-5vfnw"] Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.189670 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-qtnnc"] Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.190521 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-qtnnc" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.192968 4789 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-5wqvc" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.199060 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dkqb4"] Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.199950 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.201852 4789 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-fhfbd" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.211814 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-qtnnc"] Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.215873 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dkqb4"] Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.355690 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmp6d\" (UniqueName: \"kubernetes.io/projected/cf871183-8c76-4ab3-af84-a483ad387b7b-kube-api-access-fmp6d\") pod \"cert-manager-webhook-5655c58dd6-dkqb4\" (UID: \"cf871183-8c76-4ab3-af84-a483ad387b7b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.355845 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzsck\" (UniqueName: \"kubernetes.io/projected/a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d-kube-api-access-qzsck\") pod \"cert-manager-5b446d88c5-qtnnc\" (UID: \"a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d\") " pod="cert-manager/cert-manager-5b446d88c5-qtnnc" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.355941 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbp55\" (UniqueName: \"kubernetes.io/projected/e05b78ba-5b41-4c35-b613-87ff25afd63b-kube-api-access-qbp55\") pod \"cert-manager-cainjector-7f985d654d-5vfnw\" (UID: \"e05b78ba-5b41-4c35-b613-87ff25afd63b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-5vfnw" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.457778 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmp6d\" (UniqueName: \"kubernetes.io/projected/cf871183-8c76-4ab3-af84-a483ad387b7b-kube-api-access-fmp6d\") pod \"cert-manager-webhook-5655c58dd6-dkqb4\" (UID: \"cf871183-8c76-4ab3-af84-a483ad387b7b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.457835 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzsck\" (UniqueName: \"kubernetes.io/projected/a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d-kube-api-access-qzsck\") pod \"cert-manager-5b446d88c5-qtnnc\" (UID: \"a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d\") " pod="cert-manager/cert-manager-5b446d88c5-qtnnc" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.457862 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbp55\" (UniqueName: \"kubernetes.io/projected/e05b78ba-5b41-4c35-b613-87ff25afd63b-kube-api-access-qbp55\") pod \"cert-manager-cainjector-7f985d654d-5vfnw\" (UID: \"e05b78ba-5b41-4c35-b613-87ff25afd63b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-5vfnw" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.480554 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbp55\" (UniqueName: \"kubernetes.io/projected/e05b78ba-5b41-4c35-b613-87ff25afd63b-kube-api-access-qbp55\") pod \"cert-manager-cainjector-7f985d654d-5vfnw\" (UID: \"e05b78ba-5b41-4c35-b613-87ff25afd63b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-5vfnw" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.481896 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmp6d\" (UniqueName: \"kubernetes.io/projected/cf871183-8c76-4ab3-af84-a483ad387b7b-kube-api-access-fmp6d\") pod \"cert-manager-webhook-5655c58dd6-dkqb4\" (UID: \"cf871183-8c76-4ab3-af84-a483ad387b7b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.483082 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzsck\" (UniqueName: \"kubernetes.io/projected/a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d-kube-api-access-qzsck\") pod \"cert-manager-5b446d88c5-qtnnc\" (UID: \"a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d\") " pod="cert-manager/cert-manager-5b446d88c5-qtnnc" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.498736 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-5vfnw" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.510147 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-qtnnc" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.516513 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.708851 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-5vfnw"] Oct 08 14:11:46 crc kubenswrapper[4789]: I1008 14:11:46.714967 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 14:11:47 crc kubenswrapper[4789]: I1008 14:11:47.012703 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dkqb4"] Oct 08 14:11:47 crc kubenswrapper[4789]: I1008 14:11:47.016525 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-qtnnc"] Oct 08 14:11:47 crc kubenswrapper[4789]: W1008 14:11:47.018632 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2f9dc0d_e779_42af_a82f_e3dc45ed7b6d.slice/crio-7e492763659519ddcf8470f446018d70dd0ac98e3ce9223dd67f07571bdcbb87 WatchSource:0}: Error finding container 7e492763659519ddcf8470f446018d70dd0ac98e3ce9223dd67f07571bdcbb87: Status 404 returned error can't find the container with id 7e492763659519ddcf8470f446018d70dd0ac98e3ce9223dd67f07571bdcbb87 Oct 08 14:11:47 crc kubenswrapper[4789]: I1008 14:11:47.478838 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-5vfnw" event={"ID":"e05b78ba-5b41-4c35-b613-87ff25afd63b","Type":"ContainerStarted","Data":"7f5d09eed935d2899ffb5e97de5925c304b82b4b4ad4d2a514a51631226420f1"} Oct 08 14:11:47 crc kubenswrapper[4789]: I1008 14:11:47.479643 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-qtnnc" event={"ID":"a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d","Type":"ContainerStarted","Data":"7e492763659519ddcf8470f446018d70dd0ac98e3ce9223dd67f07571bdcbb87"} Oct 08 14:11:47 crc kubenswrapper[4789]: I1008 14:11:47.480453 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" event={"ID":"cf871183-8c76-4ab3-af84-a483ad387b7b","Type":"ContainerStarted","Data":"3a96f4e3077929e51baa13c799d751c0a793c4e47f9d918f4780a0c1aa7a5261"} Oct 08 14:11:53 crc kubenswrapper[4789]: I1008 14:11:53.521144 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-5vfnw" event={"ID":"e05b78ba-5b41-4c35-b613-87ff25afd63b","Type":"ContainerStarted","Data":"147b313c2a5310819a8dc2b998ce39617e5ff2fff88b08c5c3d2844521348ca0"} Oct 08 14:11:53 crc kubenswrapper[4789]: I1008 14:11:53.525471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-qtnnc" event={"ID":"a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d","Type":"ContainerStarted","Data":"d57dc73a9b71bf96d081c0cbd4065cfc678c85f959870b7d392ae1f30d1a3e23"} Oct 08 14:11:53 crc kubenswrapper[4789]: I1008 14:11:53.528406 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" event={"ID":"cf871183-8c76-4ab3-af84-a483ad387b7b","Type":"ContainerStarted","Data":"a9754dd9f4d16debbe721448c0b4d8f261e7313e0f71e8facb2c59fd13c92f77"} Oct 08 14:11:53 crc kubenswrapper[4789]: I1008 14:11:53.529146 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" Oct 08 14:11:53 crc kubenswrapper[4789]: I1008 14:11:53.537790 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-5vfnw" podStartSLOduration=1.990208558 podStartE2EDuration="7.537771278s" podCreationTimestamp="2025-10-08 14:11:46 +0000 UTC" firstStartedPulling="2025-10-08 14:11:46.714674721 +0000 UTC m=+646.621422213" lastFinishedPulling="2025-10-08 14:11:52.262237441 +0000 UTC m=+652.168984933" observedRunningTime="2025-10-08 14:11:53.537423389 +0000 UTC m=+653.444170891" watchObservedRunningTime="2025-10-08 14:11:53.537771278 +0000 UTC m=+653.444518760" Oct 08 14:11:53 crc kubenswrapper[4789]: I1008 14:11:53.559030 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" podStartSLOduration=1.375043429 podStartE2EDuration="7.559006694s" podCreationTimestamp="2025-10-08 14:11:46 +0000 UTC" firstStartedPulling="2025-10-08 14:11:47.018284422 +0000 UTC m=+646.925031914" lastFinishedPulling="2025-10-08 14:11:53.202247687 +0000 UTC m=+653.108995179" observedRunningTime="2025-10-08 14:11:53.55258853 +0000 UTC m=+653.459336022" watchObservedRunningTime="2025-10-08 14:11:53.559006694 +0000 UTC m=+653.465754186" Oct 08 14:11:53 crc kubenswrapper[4789]: I1008 14:11:53.571895 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-qtnnc" podStartSLOduration=2.328047217 podStartE2EDuration="7.571875953s" podCreationTimestamp="2025-10-08 14:11:46 +0000 UTC" firstStartedPulling="2025-10-08 14:11:47.020786307 +0000 UTC m=+646.927533789" lastFinishedPulling="2025-10-08 14:11:52.264615033 +0000 UTC m=+652.171362525" observedRunningTime="2025-10-08 14:11:53.568672896 +0000 UTC m=+653.475420398" watchObservedRunningTime="2025-10-08 14:11:53.571875953 +0000 UTC m=+653.478623445" Oct 08 14:11:56 crc kubenswrapper[4789]: I1008 14:11:56.508698 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9gpb2"] Oct 08 14:11:56 crc kubenswrapper[4789]: I1008 14:11:56.509828 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovn-controller" containerID="cri-o://3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a" gracePeriod=30 Oct 08 14:11:56 crc kubenswrapper[4789]: I1008 14:11:56.509940 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="nbdb" containerID="cri-o://5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c" gracePeriod=30 Oct 08 14:11:56 crc kubenswrapper[4789]: I1008 14:11:56.510051 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="northd" containerID="cri-o://2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc" gracePeriod=30 Oct 08 14:11:56 crc kubenswrapper[4789]: I1008 14:11:56.510066 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71" gracePeriod=30 Oct 08 14:11:56 crc kubenswrapper[4789]: I1008 14:11:56.510563 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kube-rbac-proxy-node" containerID="cri-o://5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be" gracePeriod=30 Oct 08 14:11:56 crc kubenswrapper[4789]: I1008 14:11:56.510645 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovn-acl-logging" containerID="cri-o://b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5" gracePeriod=30 Oct 08 14:11:56 crc kubenswrapper[4789]: I1008 14:11:56.510721 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="sbdb" containerID="cri-o://fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c" gracePeriod=30 Oct 08 14:11:56 crc kubenswrapper[4789]: I1008 14:11:56.560535 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" containerID="cri-o://135f63ea6d151315e8340b799662ffcfc90c6d9ecac7d612944bf3a97f65a626" gracePeriod=30 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.562843 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/2.log" Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.563514 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/1.log" Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.563568 4789 generic.go:334] "Generic (PLEG): container finished" podID="4aee6ee4-49ed-4135-8bb0-003eb4f70f05" containerID="b042dde4e6bd1efe56d4753229b6c7f56e3405e03b26aa92a893ee000bf1a70e" exitCode=2 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.563648 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5428v" event={"ID":"4aee6ee4-49ed-4135-8bb0-003eb4f70f05","Type":"ContainerDied","Data":"b042dde4e6bd1efe56d4753229b6c7f56e3405e03b26aa92a893ee000bf1a70e"} Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.563706 4789 scope.go:117] "RemoveContainer" containerID="25b45e3a844eafcfbf733e292ee2d9a15a25dd21737c0ef17408f54a7191d3cf" Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.564530 4789 scope.go:117] "RemoveContainer" containerID="b042dde4e6bd1efe56d4753229b6c7f56e3405e03b26aa92a893ee000bf1a70e" Oct 08 14:11:57 crc kubenswrapper[4789]: E1008 14:11:57.565423 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5428v_openshift-multus(4aee6ee4-49ed-4135-8bb0-003eb4f70f05)\"" pod="openshift-multus/multus-5428v" podUID="4aee6ee4-49ed-4135-8bb0-003eb4f70f05" Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.571459 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/3.log" Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.575136 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovn-acl-logging/0.log" Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.575731 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovn-controller/0.log" Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576275 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="135f63ea6d151315e8340b799662ffcfc90c6d9ecac7d612944bf3a97f65a626" exitCode=0 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576333 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c" exitCode=0 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576354 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c" exitCode=0 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576375 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc" exitCode=0 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576391 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71" exitCode=0 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576411 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be" exitCode=0 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576428 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5" exitCode=143 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576445 4789 generic.go:334] "Generic (PLEG): container finished" podID="1bee958d-f383-4e01-8379-b3d2ba664129" containerID="3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a" exitCode=143 Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576344 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"135f63ea6d151315e8340b799662ffcfc90c6d9ecac7d612944bf3a97f65a626"} Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576511 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c"} Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576545 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c"} Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576575 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc"} Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576599 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71"} Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576622 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be"} Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576645 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5"} Oct 08 14:11:57 crc kubenswrapper[4789]: I1008 14:11:57.576669 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a"} Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.851415 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovnkube-controller/3.log" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.852853 4789 scope.go:117] "RemoveContainer" containerID="3ca8e9d7c6960e23315f2ead705dc998e0ab4d1dbea4761750a9f16420fb2298" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.853697 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovn-acl-logging/0.log" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.854507 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-9gpb2_1bee958d-f383-4e01-8379-b3d2ba664129/ovn-controller/0.log" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.855002 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.904801 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gjmrr"] Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905098 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kube-rbac-proxy-node" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905129 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kube-rbac-proxy-node" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905151 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovn-acl-logging" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905161 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovn-acl-logging" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905174 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905186 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905203 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905213 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905225 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="nbdb" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905233 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="nbdb" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905249 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="sbdb" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905259 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="sbdb" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905272 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="northd" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905281 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="northd" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905294 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovn-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905304 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovn-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905317 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905329 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905342 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kubecfg-setup" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905352 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kubecfg-setup" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905371 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905383 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905530 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905544 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="sbdb" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905561 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovn-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905573 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905580 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovn-acl-logging" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905589 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="kube-rbac-proxy-node" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905597 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="northd" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905607 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905618 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="nbdb" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905630 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905638 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.905789 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905799 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.905917 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: E1008 14:12:00.906062 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.906072 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" containerName="ovnkube-controller" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.906729 4789 scope.go:117] "RemoveContainer" containerID="5ede30a7178405f6183d02139b5287c4dcfb815de215cd85261d70d2b1b66b4c" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.910476 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.937693 4789 scope.go:117] "RemoveContainer" containerID="fb078050d44dc6d4bf15077a5e0531fc1b4bcc694aaf8693e5b045490f1fb91c" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.958187 4789 scope.go:117] "RemoveContainer" containerID="3d1ba63191660c4bb5cddc47668e749baa353251213f6c7b4464b95bfeb8a88a" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.974854 4789 scope.go:117] "RemoveContainer" containerID="135f63ea6d151315e8340b799662ffcfc90c6d9ecac7d612944bf3a97f65a626" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978102 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-node-log\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978135 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-bin\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978175 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-systemd\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978214 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-ovn\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978249 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-log-socket\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978282 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bee958d-f383-4e01-8379-b3d2ba664129-ovn-node-metrics-cert\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978312 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-netns\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978334 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-env-overrides\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978358 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-script-lib\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978378 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-ovn-kubernetes\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978390 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-kubelet\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978407 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-systemd-units\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978421 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-etc-openvswitch\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978440 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-slash\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978458 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978477 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-config\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978492 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-openvswitch\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978511 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-netd\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978526 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zkqv\" (UniqueName: \"kubernetes.io/projected/1bee958d-f383-4e01-8379-b3d2ba664129-kube-api-access-5zkqv\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978541 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-var-lib-openvswitch\") pod \"1bee958d-f383-4e01-8379-b3d2ba664129\" (UID: \"1bee958d-f383-4e01-8379-b3d2ba664129\") " Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978665 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-kubelet\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978685 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-run-ovn\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978709 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-ovnkube-config\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978728 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-cni-netd\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978744 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-run-openvswitch\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978757 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-ovnkube-script-lib\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978773 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-run-systemd\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978790 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978806 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-systemd-units\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978825 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-env-overrides\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978845 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-cni-bin\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978862 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m8qf\" (UniqueName: \"kubernetes.io/projected/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-kube-api-access-5m8qf\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978878 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-ovn-node-metrics-cert\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978898 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-run-netns\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978920 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978940 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-log-socket\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978954 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-slash\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.978971 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-etc-openvswitch\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.979005 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-node-log\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.979027 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-var-lib-openvswitch\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.979099 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-node-log" (OuterVolumeSpecName: "node-log") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.979128 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.979640 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.979715 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-log-socket" (OuterVolumeSpecName: "log-socket") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.979866 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.979961 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.980469 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.980587 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-slash" (OuterVolumeSpecName: "host-slash") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.980623 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.980767 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.980804 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.980802 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.980834 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.980838 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.980879 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.981306 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.981489 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.994504 4789 scope.go:117] "RemoveContainer" containerID="02d98cbdbc28e60372935f50458f87991d61c92ee65db529489f74c153301d71" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.994617 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bee958d-f383-4e01-8379-b3d2ba664129-kube-api-access-5zkqv" (OuterVolumeSpecName: "kube-api-access-5zkqv") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "kube-api-access-5zkqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:12:00 crc kubenswrapper[4789]: I1008 14:12:00.995593 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.003322 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bee958d-f383-4e01-8379-b3d2ba664129-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1bee958d-f383-4e01-8379-b3d2ba664129" (UID: "1bee958d-f383-4e01-8379-b3d2ba664129"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.025576 4789 scope.go:117] "RemoveContainer" containerID="5b8fb2a34db9b808919f5dee3c392f224dc685b6618994664bb0edfc8e5c02be" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.039882 4789 scope.go:117] "RemoveContainer" containerID="2e5aa9e032aa7d8bcee72b8cf44ca6ff7815094d3f971ceeab1dfad868ddc0cc" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.058978 4789 scope.go:117] "RemoveContainer" containerID="b72d3495657067eb69f7f233b762c611b3db74ef27d4a3930b29d39c43cc43d5" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.080582 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-kubelet\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.080700 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-kubelet\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.080831 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-run-ovn\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.080691 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-run-ovn\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081146 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-ovnkube-config\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081173 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-cni-netd\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081195 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-run-openvswitch\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081212 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-ovnkube-script-lib\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081229 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-run-systemd\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081247 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081265 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-systemd-units\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081281 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-env-overrides\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081302 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-cni-bin\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081317 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m8qf\" (UniqueName: \"kubernetes.io/projected/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-kube-api-access-5m8qf\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081336 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-ovn-node-metrics-cert\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081363 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-run-netns\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081390 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081416 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-slash\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081433 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-log-socket\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081454 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-etc-openvswitch\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081469 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-node-log\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081499 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-var-lib-openvswitch\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081546 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081560 4789 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081571 4789 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081582 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zkqv\" (UniqueName: \"kubernetes.io/projected/1bee958d-f383-4e01-8379-b3d2ba664129-kube-api-access-5zkqv\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081592 4789 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081602 4789 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-node-log\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081612 4789 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081623 4789 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081634 4789 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081645 4789 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-log-socket\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081655 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1bee958d-f383-4e01-8379-b3d2ba664129-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081668 4789 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081679 4789 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081691 4789 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1bee958d-f383-4e01-8379-b3d2ba664129-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081703 4789 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081714 4789 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081725 4789 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081736 4789 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081746 4789 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-slash\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081757 4789 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bee958d-f383-4e01-8379-b3d2ba664129-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.081791 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-var-lib-openvswitch\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.083927 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-cni-netd\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.083962 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-run-openvswitch\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084443 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-ovnkube-script-lib\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084480 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-run-netns\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084501 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-etc-openvswitch\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084518 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-node-log\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084542 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084562 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-slash\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084581 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-systemd-units\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084576 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-log-socket\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084615 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084637 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-host-cni-bin\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084599 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-run-systemd\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.084903 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-ovnkube-config\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.085295 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-env-overrides\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.089132 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-ovn-node-metrics-cert\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.091509 4789 scope.go:117] "RemoveContainer" containerID="75dd732c3a1b1aa1208e3c7526be17845dd80021faf01c3645ab68ca96cfa66f" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.102949 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m8qf\" (UniqueName: \"kubernetes.io/projected/a578cdd5-acd6-4325-8ca1-e0fc95b480a1-kube-api-access-5m8qf\") pod \"ovnkube-node-gjmrr\" (UID: \"a578cdd5-acd6-4325-8ca1-e0fc95b480a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.246297 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:01 crc kubenswrapper[4789]: W1008 14:12:01.263174 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda578cdd5_acd6_4325_8ca1_e0fc95b480a1.slice/crio-d130ec269372e3f40d94822bd60c1a23467b7b2404aa04b88fe411af3bc49f57 WatchSource:0}: Error finding container d130ec269372e3f40d94822bd60c1a23467b7b2404aa04b88fe411af3bc49f57: Status 404 returned error can't find the container with id d130ec269372e3f40d94822bd60c1a23467b7b2404aa04b88fe411af3bc49f57 Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.519736 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-dkqb4" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.604738 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/2.log" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.606451 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerStarted","Data":"d130ec269372e3f40d94822bd60c1a23467b7b2404aa04b88fe411af3bc49f57"} Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.606505 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.606509 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9gpb2" event={"ID":"1bee958d-f383-4e01-8379-b3d2ba664129","Type":"ContainerDied","Data":"eb0c8cc80e98c32256c7417583d52a14a2c243b48d61b6c50ca9ec08f6a05811"} Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.633879 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9gpb2"] Oct 08 14:12:01 crc kubenswrapper[4789]: I1008 14:12:01.643079 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9gpb2"] Oct 08 14:12:02 crc kubenswrapper[4789]: I1008 14:12:02.625269 4789 generic.go:334] "Generic (PLEG): container finished" podID="a578cdd5-acd6-4325-8ca1-e0fc95b480a1" containerID="ae64270009936da5b351e21ed46113e7941e2c00b5b6940febc25b0391d56f9f" exitCode=0 Oct 08 14:12:02 crc kubenswrapper[4789]: I1008 14:12:02.625625 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerDied","Data":"ae64270009936da5b351e21ed46113e7941e2c00b5b6940febc25b0391d56f9f"} Oct 08 14:12:02 crc kubenswrapper[4789]: I1008 14:12:02.741525 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bee958d-f383-4e01-8379-b3d2ba664129" path="/var/lib/kubelet/pods/1bee958d-f383-4e01-8379-b3d2ba664129/volumes" Oct 08 14:12:03 crc kubenswrapper[4789]: I1008 14:12:03.634514 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerStarted","Data":"4f543a703957436059533e9236415a6bb44fa2fe74cb92474f96345083471387"} Oct 08 14:12:04 crc kubenswrapper[4789]: I1008 14:12:04.642392 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerStarted","Data":"cbf500c83cada69bb873aceec614e4bc47f9ab315de84f81682e623ad5ed4bf0"} Oct 08 14:12:04 crc kubenswrapper[4789]: I1008 14:12:04.642764 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerStarted","Data":"765a8ca0fe0d318ab062a1bd198c9a193133daedf6f5f2efd7382a12603886fe"} Oct 08 14:12:05 crc kubenswrapper[4789]: I1008 14:12:05.650764 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerStarted","Data":"d1a877d73f09953b0aca255f6162a2eb945ce9392146745d745146eaff6ca153"} Oct 08 14:12:06 crc kubenswrapper[4789]: I1008 14:12:06.663669 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerStarted","Data":"09016cec1afc5eeb9eddead589e937114f3c6463337d045baf6ccdd30e9fe1b1"} Oct 08 14:12:07 crc kubenswrapper[4789]: I1008 14:12:07.673498 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerStarted","Data":"0c1bf7cf7936d0255b6681ed726b21f5e0508302a0117d38e4a0ace78c76fafa"} Oct 08 14:12:08 crc kubenswrapper[4789]: I1008 14:12:08.689734 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerStarted","Data":"fd7d6a4a9b1387372ce9140a4a287dc6191f0de0c0eaca9bb39dffc3fd587708"} Oct 08 14:12:10 crc kubenswrapper[4789]: I1008 14:12:10.732835 4789 scope.go:117] "RemoveContainer" containerID="b042dde4e6bd1efe56d4753229b6c7f56e3405e03b26aa92a893ee000bf1a70e" Oct 08 14:12:10 crc kubenswrapper[4789]: E1008 14:12:10.733120 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5428v_openshift-multus(4aee6ee4-49ed-4135-8bb0-003eb4f70f05)\"" pod="openshift-multus/multus-5428v" podUID="4aee6ee4-49ed-4135-8bb0-003eb4f70f05" Oct 08 14:12:11 crc kubenswrapper[4789]: I1008 14:12:11.715163 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" event={"ID":"a578cdd5-acd6-4325-8ca1-e0fc95b480a1","Type":"ContainerStarted","Data":"04af440bd4036921c9e811c0c01eadceef22505d87ba036154fd16a76880a58f"} Oct 08 14:12:11 crc kubenswrapper[4789]: I1008 14:12:11.715798 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:11 crc kubenswrapper[4789]: I1008 14:12:11.715825 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:11 crc kubenswrapper[4789]: I1008 14:12:11.715844 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:11 crc kubenswrapper[4789]: I1008 14:12:11.747796 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:11 crc kubenswrapper[4789]: I1008 14:12:11.747920 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:11 crc kubenswrapper[4789]: I1008 14:12:11.754057 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" podStartSLOduration=11.754042249 podStartE2EDuration="11.754042249s" podCreationTimestamp="2025-10-08 14:12:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:12:11.752662842 +0000 UTC m=+671.659410334" watchObservedRunningTime="2025-10-08 14:12:11.754042249 +0000 UTC m=+671.660789741" Oct 08 14:12:22 crc kubenswrapper[4789]: I1008 14:12:22.729660 4789 scope.go:117] "RemoveContainer" containerID="b042dde4e6bd1efe56d4753229b6c7f56e3405e03b26aa92a893ee000bf1a70e" Oct 08 14:12:23 crc kubenswrapper[4789]: I1008 14:12:23.785755 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5428v_4aee6ee4-49ed-4135-8bb0-003eb4f70f05/kube-multus/2.log" Oct 08 14:12:23 crc kubenswrapper[4789]: I1008 14:12:23.786097 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5428v" event={"ID":"4aee6ee4-49ed-4135-8bb0-003eb4f70f05","Type":"ContainerStarted","Data":"d9a54cb82298345c47c48c66299399a71fd61172a2fce7148d18f5abb7ee9c74"} Oct 08 14:12:31 crc kubenswrapper[4789]: I1008 14:12:31.266813 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gjmrr" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.095295 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn"] Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.097694 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.100391 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.113374 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn"] Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.177618 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.177931 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdmcj\" (UniqueName: \"kubernetes.io/projected/fa1275e9-3700-47b0-b09f-a68a046cc1e4-kube-api-access-kdmcj\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.178103 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.279646 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.279741 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.279779 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdmcj\" (UniqueName: \"kubernetes.io/projected/fa1275e9-3700-47b0-b09f-a68a046cc1e4-kube-api-access-kdmcj\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.280293 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.280360 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.297686 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdmcj\" (UniqueName: \"kubernetes.io/projected/fa1275e9-3700-47b0-b09f-a68a046cc1e4-kube-api-access-kdmcj\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.419457 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.614399 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn"] Oct 08 14:12:36 crc kubenswrapper[4789]: I1008 14:12:36.852342 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" event={"ID":"fa1275e9-3700-47b0-b09f-a68a046cc1e4","Type":"ContainerStarted","Data":"cb823e133e606f58d1ae2329faff30e62479ac7c0e896cbdf542c5b95991b356"} Oct 08 14:12:37 crc kubenswrapper[4789]: I1008 14:12:37.859136 4789 generic.go:334] "Generic (PLEG): container finished" podID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerID="df08cf5609fb6f4647a8874ffd87609f7c89dc3b1e0a4cc4e8feb7f309287ba9" exitCode=0 Oct 08 14:12:37 crc kubenswrapper[4789]: I1008 14:12:37.859457 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" event={"ID":"fa1275e9-3700-47b0-b09f-a68a046cc1e4","Type":"ContainerDied","Data":"df08cf5609fb6f4647a8874ffd87609f7c89dc3b1e0a4cc4e8feb7f309287ba9"} Oct 08 14:12:56 crc kubenswrapper[4789]: I1008 14:12:56.433124 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:12:56 crc kubenswrapper[4789]: I1008 14:12:56.433857 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:12:56 crc kubenswrapper[4789]: I1008 14:12:56.980900 4789 generic.go:334] "Generic (PLEG): container finished" podID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerID="085df2c7417d3145e68c54f57272eb166f5bb142638f396927dca7d1833ca6f9" exitCode=0 Oct 08 14:12:56 crc kubenswrapper[4789]: I1008 14:12:56.980938 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" event={"ID":"fa1275e9-3700-47b0-b09f-a68a046cc1e4","Type":"ContainerDied","Data":"085df2c7417d3145e68c54f57272eb166f5bb142638f396927dca7d1833ca6f9"} Oct 08 14:12:57 crc kubenswrapper[4789]: I1008 14:12:57.988448 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" event={"ID":"fa1275e9-3700-47b0-b09f-a68a046cc1e4","Type":"ContainerStarted","Data":"71fb18fe8a31d859742e25c9414cafbb1e442afd1b4257cde06e361d00190591"} Oct 08 14:12:58 crc kubenswrapper[4789]: I1008 14:12:58.005552 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" podStartSLOduration=4.029955236 podStartE2EDuration="22.005535029s" podCreationTimestamp="2025-10-08 14:12:36 +0000 UTC" firstStartedPulling="2025-10-08 14:12:37.861821103 +0000 UTC m=+697.768568605" lastFinishedPulling="2025-10-08 14:12:55.837400896 +0000 UTC m=+715.744148398" observedRunningTime="2025-10-08 14:12:58.003794111 +0000 UTC m=+717.910541613" watchObservedRunningTime="2025-10-08 14:12:58.005535029 +0000 UTC m=+717.912282521" Oct 08 14:12:58 crc kubenswrapper[4789]: I1008 14:12:58.995078 4789 generic.go:334] "Generic (PLEG): container finished" podID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerID="71fb18fe8a31d859742e25c9414cafbb1e442afd1b4257cde06e361d00190591" exitCode=0 Oct 08 14:12:58 crc kubenswrapper[4789]: I1008 14:12:58.995172 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" event={"ID":"fa1275e9-3700-47b0-b09f-a68a046cc1e4","Type":"ContainerDied","Data":"71fb18fe8a31d859742e25c9414cafbb1e442afd1b4257cde06e361d00190591"} Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.270353 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.402622 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdmcj\" (UniqueName: \"kubernetes.io/projected/fa1275e9-3700-47b0-b09f-a68a046cc1e4-kube-api-access-kdmcj\") pod \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.403201 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-util\") pod \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.403351 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-bundle\") pod \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\" (UID: \"fa1275e9-3700-47b0-b09f-a68a046cc1e4\") " Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.406978 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-bundle" (OuterVolumeSpecName: "bundle") pod "fa1275e9-3700-47b0-b09f-a68a046cc1e4" (UID: "fa1275e9-3700-47b0-b09f-a68a046cc1e4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.410884 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa1275e9-3700-47b0-b09f-a68a046cc1e4-kube-api-access-kdmcj" (OuterVolumeSpecName: "kube-api-access-kdmcj") pod "fa1275e9-3700-47b0-b09f-a68a046cc1e4" (UID: "fa1275e9-3700-47b0-b09f-a68a046cc1e4"). InnerVolumeSpecName "kube-api-access-kdmcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.413730 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-util" (OuterVolumeSpecName: "util") pod "fa1275e9-3700-47b0-b09f-a68a046cc1e4" (UID: "fa1275e9-3700-47b0-b09f-a68a046cc1e4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.506480 4789 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.506545 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdmcj\" (UniqueName: \"kubernetes.io/projected/fa1275e9-3700-47b0-b09f-a68a046cc1e4-kube-api-access-kdmcj\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:00 crc kubenswrapper[4789]: I1008 14:13:00.506568 4789 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa1275e9-3700-47b0-b09f-a68a046cc1e4-util\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:01 crc kubenswrapper[4789]: I1008 14:13:01.006522 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" event={"ID":"fa1275e9-3700-47b0-b09f-a68a046cc1e4","Type":"ContainerDied","Data":"cb823e133e606f58d1ae2329faff30e62479ac7c0e896cbdf542c5b95991b356"} Oct 08 14:13:01 crc kubenswrapper[4789]: I1008 14:13:01.006575 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn" Oct 08 14:13:01 crc kubenswrapper[4789]: I1008 14:13:01.006569 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb823e133e606f58d1ae2329faff30e62479ac7c0e896cbdf542c5b95991b356" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.208579 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt"] Oct 08 14:13:08 crc kubenswrapper[4789]: E1008 14:13:08.209380 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerName="pull" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.209396 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerName="pull" Oct 08 14:13:08 crc kubenswrapper[4789]: E1008 14:13:08.209410 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerName="util" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.209417 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerName="util" Oct 08 14:13:08 crc kubenswrapper[4789]: E1008 14:13:08.209438 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerName="extract" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.209447 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerName="extract" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.209557 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa1275e9-3700-47b0-b09f-a68a046cc1e4" containerName="extract" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.210037 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.212811 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.216021 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.216272 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-zlqjl" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.248850 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.249654 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.259328 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.262241 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.262474 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-btxqq" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.271892 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.272717 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.275145 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.290057 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.302054 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhsh2\" (UniqueName: \"kubernetes.io/projected/5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec-kube-api-access-zhsh2\") pod \"obo-prometheus-operator-7c8cf85677-lwgzt\" (UID: \"5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.403066 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7141b3e4-ea4a-4476-84d2-57670fe5e932-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh\" (UID: \"7141b3e4-ea4a-4476-84d2-57670fe5e932\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.403128 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhsh2\" (UniqueName: \"kubernetes.io/projected/5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec-kube-api-access-zhsh2\") pod \"obo-prometheus-operator-7c8cf85677-lwgzt\" (UID: \"5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.403155 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7141b3e4-ea4a-4476-84d2-57670fe5e932-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh\" (UID: \"7141b3e4-ea4a-4476-84d2-57670fe5e932\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.403355 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9451e6d9-fa5b-4209-a332-a65412994d55-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l\" (UID: \"9451e6d9-fa5b-4209-a332-a65412994d55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.403434 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9451e6d9-fa5b-4209-a332-a65412994d55-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l\" (UID: \"9451e6d9-fa5b-4209-a332-a65412994d55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.422527 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhsh2\" (UniqueName: \"kubernetes.io/projected/5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec-kube-api-access-zhsh2\") pod \"obo-prometheus-operator-7c8cf85677-lwgzt\" (UID: \"5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.449870 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-tclgl"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.450706 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.454629 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.454642 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-v2nch" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.470046 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-tclgl"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.509512 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9451e6d9-fa5b-4209-a332-a65412994d55-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l\" (UID: \"9451e6d9-fa5b-4209-a332-a65412994d55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.509583 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7141b3e4-ea4a-4476-84d2-57670fe5e932-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh\" (UID: \"7141b3e4-ea4a-4476-84d2-57670fe5e932\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.509613 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7141b3e4-ea4a-4476-84d2-57670fe5e932-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh\" (UID: \"7141b3e4-ea4a-4476-84d2-57670fe5e932\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.509655 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9451e6d9-fa5b-4209-a332-a65412994d55-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l\" (UID: \"9451e6d9-fa5b-4209-a332-a65412994d55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.512936 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9451e6d9-fa5b-4209-a332-a65412994d55-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l\" (UID: \"9451e6d9-fa5b-4209-a332-a65412994d55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.513902 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9451e6d9-fa5b-4209-a332-a65412994d55-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l\" (UID: \"9451e6d9-fa5b-4209-a332-a65412994d55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.516117 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7141b3e4-ea4a-4476-84d2-57670fe5e932-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh\" (UID: \"7141b3e4-ea4a-4476-84d2-57670fe5e932\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.520739 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7141b3e4-ea4a-4476-84d2-57670fe5e932-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh\" (UID: \"7141b3e4-ea4a-4476-84d2-57670fe5e932\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.528682 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.564140 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.579223 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-rbwl7"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.579943 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.584109 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-5k7rs" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.590754 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.591110 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-rbwl7"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.612008 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/643a32e3-75dd-44f5-bbac-e2d917641a29-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-tclgl\" (UID: \"643a32e3-75dd-44f5-bbac-e2d917641a29\") " pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.612242 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sgwc\" (UniqueName: \"kubernetes.io/projected/643a32e3-75dd-44f5-bbac-e2d917641a29-kube-api-access-9sgwc\") pod \"observability-operator-cc5f78dfc-tclgl\" (UID: \"643a32e3-75dd-44f5-bbac-e2d917641a29\") " pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.714002 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-rbwl7\" (UID: \"1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5\") " pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.714353 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sgwc\" (UniqueName: \"kubernetes.io/projected/643a32e3-75dd-44f5-bbac-e2d917641a29-kube-api-access-9sgwc\") pod \"observability-operator-cc5f78dfc-tclgl\" (UID: \"643a32e3-75dd-44f5-bbac-e2d917641a29\") " pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.714372 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwljk\" (UniqueName: \"kubernetes.io/projected/1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5-kube-api-access-qwljk\") pod \"perses-operator-54bc95c9fb-rbwl7\" (UID: \"1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5\") " pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.714404 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/643a32e3-75dd-44f5-bbac-e2d917641a29-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-tclgl\" (UID: \"643a32e3-75dd-44f5-bbac-e2d917641a29\") " pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.725863 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/643a32e3-75dd-44f5-bbac-e2d917641a29-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-tclgl\" (UID: \"643a32e3-75dd-44f5-bbac-e2d917641a29\") " pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.739857 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sgwc\" (UniqueName: \"kubernetes.io/projected/643a32e3-75dd-44f5-bbac-e2d917641a29-kube-api-access-9sgwc\") pod \"observability-operator-cc5f78dfc-tclgl\" (UID: \"643a32e3-75dd-44f5-bbac-e2d917641a29\") " pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.768232 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.815751 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-rbwl7\" (UID: \"1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5\") " pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.815820 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwljk\" (UniqueName: \"kubernetes.io/projected/1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5-kube-api-access-qwljk\") pod \"perses-operator-54bc95c9fb-rbwl7\" (UID: \"1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5\") " pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.816671 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-rbwl7\" (UID: \"1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5\") " pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.842122 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwljk\" (UniqueName: \"kubernetes.io/projected/1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5-kube-api-access-qwljk\") pod \"perses-operator-54bc95c9fb-rbwl7\" (UID: \"1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5\") " pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.903058 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh"] Oct 08 14:13:08 crc kubenswrapper[4789]: I1008 14:13:08.916539 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:09 crc kubenswrapper[4789]: I1008 14:13:09.051750 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt"] Oct 08 14:13:09 crc kubenswrapper[4789]: I1008 14:13:09.082838 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" event={"ID":"7141b3e4-ea4a-4476-84d2-57670fe5e932","Type":"ContainerStarted","Data":"4ae6f8dee8e5ec48830beffad74fedafb7349c62c7c584b2598c9ff27c05688d"} Oct 08 14:13:09 crc kubenswrapper[4789]: I1008 14:13:09.087612 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-tclgl"] Oct 08 14:13:09 crc kubenswrapper[4789]: W1008 14:13:09.098266 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f7cb3da_f7e4_49ec_9ad4_dea3650cf5ec.slice/crio-424033bdb80b40776cc2cfcc1369beca5d0ac0f5e77aa35e568ca77562d2a8e3 WatchSource:0}: Error finding container 424033bdb80b40776cc2cfcc1369beca5d0ac0f5e77aa35e568ca77562d2a8e3: Status 404 returned error can't find the container with id 424033bdb80b40776cc2cfcc1369beca5d0ac0f5e77aa35e568ca77562d2a8e3 Oct 08 14:13:09 crc kubenswrapper[4789]: I1008 14:13:09.223377 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l"] Oct 08 14:13:09 crc kubenswrapper[4789]: I1008 14:13:09.263721 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-rbwl7"] Oct 08 14:13:09 crc kubenswrapper[4789]: W1008 14:13:09.267246 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cc6e22a_f412_459c_a8b4_5ee1fabe0bd5.slice/crio-5b360f0cceced96224da65769f1a76bf6757687e6d6c8a97435a268453f94394 WatchSource:0}: Error finding container 5b360f0cceced96224da65769f1a76bf6757687e6d6c8a97435a268453f94394: Status 404 returned error can't find the container with id 5b360f0cceced96224da65769f1a76bf6757687e6d6c8a97435a268453f94394 Oct 08 14:13:10 crc kubenswrapper[4789]: I1008 14:13:10.088223 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" event={"ID":"1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5","Type":"ContainerStarted","Data":"5b360f0cceced96224da65769f1a76bf6757687e6d6c8a97435a268453f94394"} Oct 08 14:13:10 crc kubenswrapper[4789]: I1008 14:13:10.090321 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" event={"ID":"5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec","Type":"ContainerStarted","Data":"424033bdb80b40776cc2cfcc1369beca5d0ac0f5e77aa35e568ca77562d2a8e3"} Oct 08 14:13:10 crc kubenswrapper[4789]: I1008 14:13:10.091700 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" event={"ID":"643a32e3-75dd-44f5-bbac-e2d917641a29","Type":"ContainerStarted","Data":"d744e00eda20db4c72f9b09ed46c104376d7045041e8854265beb8d5a3c8a34f"} Oct 08 14:13:10 crc kubenswrapper[4789]: I1008 14:13:10.092883 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" event={"ID":"9451e6d9-fa5b-4209-a332-a65412994d55","Type":"ContainerStarted","Data":"ce5071094d90927e9950c450e0f6325e20865737ed21a8c3d1edaff5dc28ad0a"} Oct 08 14:13:24 crc kubenswrapper[4789]: E1008 14:13:24.625333 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5" Oct 08 14:13:24 crc kubenswrapper[4789]: E1008 14:13:24.626046 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5,Command:[],Args:[--prometheus-config-reloader=$(RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER) --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOGC,Value:30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER,Value:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:8597c48fc71fc6ec8e87dbe40dace4dbb7b817c1039db608af76a0d90f7ac2d0,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{157286400 0} {} 150Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhsh2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-7c8cf85677-lwgzt_openshift-operators(5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:13:24 crc kubenswrapper[4789]: E1008 14:13:24.627231 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" podUID="5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec" Oct 08 14:13:24 crc kubenswrapper[4789]: E1008 14:13:24.964045 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c" Oct 08 14:13:24 crc kubenswrapper[4789]: E1008 14:13:24.964221 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwljk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-54bc95c9fb-rbwl7_openshift-operators(1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:13:24 crc kubenswrapper[4789]: E1008 14:13:24.965706 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" podUID="1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5" Oct 08 14:13:25 crc kubenswrapper[4789]: E1008 14:13:25.193749 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e2681bce57dc9c15701f5591532c2dfe8f19778606661339553a28dc003dbca5\\\"\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" podUID="5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec" Oct 08 14:13:25 crc kubenswrapper[4789]: E1008 14:13:25.193784 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c\\\"\"" pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" podUID="1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5" Oct 08 14:13:25 crc kubenswrapper[4789]: E1008 14:13:25.579965 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d" Oct 08 14:13:25 crc kubenswrapper[4789]: E1008 14:13:25.580545 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l_openshift-operators(9451e6d9-fa5b-4209-a332-a65412994d55): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:13:25 crc kubenswrapper[4789]: E1008 14:13:25.581785 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" podUID="9451e6d9-fa5b-4209-a332-a65412994d55" Oct 08 14:13:25 crc kubenswrapper[4789]: E1008 14:13:25.627909 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d" Oct 08 14:13:25 crc kubenswrapper[4789]: E1008 14:13:25.628123 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh_openshift-operators(7141b3e4-ea4a-4476-84d2-57670fe5e932): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:13:25 crc kubenswrapper[4789]: E1008 14:13:25.629533 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" podUID="7141b3e4-ea4a-4476-84d2-57670fe5e932" Oct 08 14:13:26 crc kubenswrapper[4789]: E1008 14:13:26.197898 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" podUID="9451e6d9-fa5b-4209-a332-a65412994d55" Oct 08 14:13:26 crc kubenswrapper[4789]: E1008 14:13:26.200042 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:e54c1e1301be66933f3ecb01d5a0ca27f58aabfd905b18b7d057bbf23bdb7b0d\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" podUID="7141b3e4-ea4a-4476-84d2-57670fe5e932" Oct 08 14:13:26 crc kubenswrapper[4789]: I1008 14:13:26.433361 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:13:26 crc kubenswrapper[4789]: I1008 14:13:26.433418 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:13:27 crc kubenswrapper[4789]: E1008 14:13:27.763254 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e" Oct 08 14:13:27 crc kubenswrapper[4789]: E1008 14:13:27.763771 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e,Command:[],Args:[--namespace=$(NAMESPACE) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=perses=$(RELATED_IMAGE_PERSES) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:4d25b0e31549d780928d2dd3eed7defd9c6d460deb92dcff0fe72c5023029404,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:a0a1d0e39de54c5b2786c2b82d0104f358b479135c069075ddd4f7cd76826c00,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:f3806c97420ec8ba91895ce7627df7612cccb927c05d7854377f45cdd6c924a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-0-50-rhel9@sha256:4b5e53d226733237fc5abd0476eb3c96162cf3d8da7aeba8deda631fa8987223,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-0-4-rhel9@sha256:53125bddbefca2ba2b57c3fd74bd4b376da803e420201220548878f557bd6610,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-1-0-rhel9@sha256:1dbe9a684271e00c8f36d8b96c9b22f6ee3c6f907ea6ad20980901bd533f9a3a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-4-rhel9@sha256:6aafab2c90bcbc6702f2d63d585a764baa8de8207e6af7afa60f3976ddfa9bd3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-0-3-rhel9@sha256:9f80851e8137c2c5e5c2aee13fc663f6c7124d9524d88c06c1507748ce84e1ed,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-1-rhel9@sha256:2c9b2be12f15f06a24393dbab6a31682cee399d42e2cc04b0dcf03b2b598d5cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-6-0-rhel9@sha256:e9042d93f624790c450724158a8323277e4dd136530c763fec8db31f51fd8552,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-0-4-rhel9@sha256:456d45001816b9adc38745e0ad8705bdc0150d03d0f65e0dfa9caf3fb8980fad,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-5-rhel9@sha256:f3446969c67c18b44bee38ac946091fe9397a2117cb5b7aacb39406461c1efe1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-0-4-rhel9@sha256:ade84f8be7d23bd4b9c80e07462dc947280f0bcf6071e6edd927fef54c254b7e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:039e139cf9217bbe72248674df76cbe4baf4bef9f8dc367d2cb51eae9c4aa9d7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:142180f277f0221ef2d4176f9af6dcdb4e7ab434a68f0dfad2ee5bee0e667ddd,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9sgwc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-cc5f78dfc-tclgl_openshift-operators(643a32e3-75dd-44f5-bbac-e2d917641a29): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 14:13:27 crc kubenswrapper[4789]: E1008 14:13:27.764943 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" podUID="643a32e3-75dd-44f5-bbac-e2d917641a29" Oct 08 14:13:28 crc kubenswrapper[4789]: E1008 14:13:28.209257 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:27ffe36aad6e606e6d0a211f48f3cdb58a53aa0d5e8ead6a444427231261ab9e\\\"\"" pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" podUID="643a32e3-75dd-44f5-bbac-e2d917641a29" Oct 08 14:13:28 crc kubenswrapper[4789]: I1008 14:13:28.638881 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rhnqk"] Oct 08 14:13:28 crc kubenswrapper[4789]: I1008 14:13:28.639354 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" podUID="a9100571-6717-43fe-a0d8-0da2bd0ce31c" containerName="controller-manager" containerID="cri-o://6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5" gracePeriod=30 Oct 08 14:13:28 crc kubenswrapper[4789]: I1008 14:13:28.655175 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d"] Oct 08 14:13:28 crc kubenswrapper[4789]: I1008 14:13:28.655576 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" podUID="5ae1d9aa-857d-416d-92a1-e08e45bd7b19" containerName="route-controller-manager" containerID="cri-o://61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22" gracePeriod=30 Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.135176 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.167870 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.213364 4789 generic.go:334] "Generic (PLEG): container finished" podID="a9100571-6717-43fe-a0d8-0da2bd0ce31c" containerID="6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5" exitCode=0 Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.213420 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" event={"ID":"a9100571-6717-43fe-a0d8-0da2bd0ce31c","Type":"ContainerDied","Data":"6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5"} Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.213445 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" event={"ID":"a9100571-6717-43fe-a0d8-0da2bd0ce31c","Type":"ContainerDied","Data":"54be8b8f18e145300894a81e3624eb13ca8a917defdee2c30c3708705bd12c4f"} Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.213461 4789 scope.go:117] "RemoveContainer" containerID="6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.213555 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rhnqk" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.220026 4789 generic.go:334] "Generic (PLEG): container finished" podID="5ae1d9aa-857d-416d-92a1-e08e45bd7b19" containerID="61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22" exitCode=0 Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.220055 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" event={"ID":"5ae1d9aa-857d-416d-92a1-e08e45bd7b19","Type":"ContainerDied","Data":"61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22"} Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.220249 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" event={"ID":"5ae1d9aa-857d-416d-92a1-e08e45bd7b19","Type":"ContainerDied","Data":"52a12cd8b8724a5076c9d4e6809ba6c6d2ada7c85022c0b173dc5f309d74fe4d"} Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.220290 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.236451 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-client-ca\") pod \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.236507 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbzwl\" (UniqueName: \"kubernetes.io/projected/a9100571-6717-43fe-a0d8-0da2bd0ce31c-kube-api-access-xbzwl\") pod \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.236557 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-config\") pod \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.236581 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-proxy-ca-bundles\") pod \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.236605 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-serving-cert\") pod \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.236672 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-config\") pod \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.236695 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-client-ca\") pod \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.236729 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6dgm\" (UniqueName: \"kubernetes.io/projected/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-kube-api-access-l6dgm\") pod \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\" (UID: \"5ae1d9aa-857d-416d-92a1-e08e45bd7b19\") " Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.236794 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9100571-6717-43fe-a0d8-0da2bd0ce31c-serving-cert\") pod \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\" (UID: \"a9100571-6717-43fe-a0d8-0da2bd0ce31c\") " Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.238072 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-client-ca" (OuterVolumeSpecName: "client-ca") pod "a9100571-6717-43fe-a0d8-0da2bd0ce31c" (UID: "a9100571-6717-43fe-a0d8-0da2bd0ce31c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.239869 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-config" (OuterVolumeSpecName: "config") pod "5ae1d9aa-857d-416d-92a1-e08e45bd7b19" (UID: "5ae1d9aa-857d-416d-92a1-e08e45bd7b19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.240535 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a9100571-6717-43fe-a0d8-0da2bd0ce31c" (UID: "a9100571-6717-43fe-a0d8-0da2bd0ce31c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.241357 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-config" (OuterVolumeSpecName: "config") pod "a9100571-6717-43fe-a0d8-0da2bd0ce31c" (UID: "a9100571-6717-43fe-a0d8-0da2bd0ce31c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.241454 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-client-ca" (OuterVolumeSpecName: "client-ca") pod "5ae1d9aa-857d-416d-92a1-e08e45bd7b19" (UID: "5ae1d9aa-857d-416d-92a1-e08e45bd7b19"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.244295 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9100571-6717-43fe-a0d8-0da2bd0ce31c-kube-api-access-xbzwl" (OuterVolumeSpecName: "kube-api-access-xbzwl") pod "a9100571-6717-43fe-a0d8-0da2bd0ce31c" (UID: "a9100571-6717-43fe-a0d8-0da2bd0ce31c"). InnerVolumeSpecName "kube-api-access-xbzwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.248161 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9100571-6717-43fe-a0d8-0da2bd0ce31c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a9100571-6717-43fe-a0d8-0da2bd0ce31c" (UID: "a9100571-6717-43fe-a0d8-0da2bd0ce31c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.248924 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5ae1d9aa-857d-416d-92a1-e08e45bd7b19" (UID: "5ae1d9aa-857d-416d-92a1-e08e45bd7b19"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.255240 4789 scope.go:117] "RemoveContainer" containerID="6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5" Oct 08 14:13:29 crc kubenswrapper[4789]: E1008 14:13:29.255673 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5\": container with ID starting with 6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5 not found: ID does not exist" containerID="6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.255711 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5"} err="failed to get container status \"6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5\": rpc error: code = NotFound desc = could not find container \"6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5\": container with ID starting with 6147fff7cd7ee9b4129141fbab2e0b770d4df2bd2b54caa4f17fde90f8e81ec5 not found: ID does not exist" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.255737 4789 scope.go:117] "RemoveContainer" containerID="61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.255796 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-kube-api-access-l6dgm" (OuterVolumeSpecName: "kube-api-access-l6dgm") pod "5ae1d9aa-857d-416d-92a1-e08e45bd7b19" (UID: "5ae1d9aa-857d-416d-92a1-e08e45bd7b19"). InnerVolumeSpecName "kube-api-access-l6dgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.273323 4789 scope.go:117] "RemoveContainer" containerID="61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22" Oct 08 14:13:29 crc kubenswrapper[4789]: E1008 14:13:29.273781 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22\": container with ID starting with 61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22 not found: ID does not exist" containerID="61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.273812 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22"} err="failed to get container status \"61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22\": rpc error: code = NotFound desc = could not find container \"61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22\": container with ID starting with 61895fc159f3f00d9714282426405bed8c269614f6956a8cd1f34ab36686cb22 not found: ID does not exist" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.337802 4789 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.337837 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6dgm\" (UniqueName: \"kubernetes.io/projected/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-kube-api-access-l6dgm\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.337848 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9100571-6717-43fe-a0d8-0da2bd0ce31c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.337859 4789 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.337868 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbzwl\" (UniqueName: \"kubernetes.io/projected/a9100571-6717-43fe-a0d8-0da2bd0ce31c-kube-api-access-xbzwl\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.337876 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.337886 4789 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.337897 4789 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae1d9aa-857d-416d-92a1-e08e45bd7b19-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.337905 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9100571-6717-43fe-a0d8-0da2bd0ce31c-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.551287 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rhnqk"] Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.556904 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rhnqk"] Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.580006 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d"] Oct 08 14:13:29 crc kubenswrapper[4789]: I1008 14:13:29.584202 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7gr6d"] Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.582014 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd"] Oct 08 14:13:30 crc kubenswrapper[4789]: E1008 14:13:30.582594 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae1d9aa-857d-416d-92a1-e08e45bd7b19" containerName="route-controller-manager" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.582609 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae1d9aa-857d-416d-92a1-e08e45bd7b19" containerName="route-controller-manager" Oct 08 14:13:30 crc kubenswrapper[4789]: E1008 14:13:30.582623 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9100571-6717-43fe-a0d8-0da2bd0ce31c" containerName="controller-manager" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.582632 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9100571-6717-43fe-a0d8-0da2bd0ce31c" containerName="controller-manager" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.582752 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae1d9aa-857d-416d-92a1-e08e45bd7b19" containerName="route-controller-manager" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.582771 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9100571-6717-43fe-a0d8-0da2bd0ce31c" containerName="controller-manager" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.583271 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.587266 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v"] Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.588710 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.595329 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.598353 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.598730 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.601551 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.602246 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.602684 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.603664 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.604893 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.604920 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.605029 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.605077 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.605028 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.606660 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd"] Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.608470 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.611809 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v"] Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.653194 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-client-ca\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.653236 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-serving-cert\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.653258 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-config\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.653278 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c114d60-0f48-48f1-ba2f-e9c3e070486f-client-ca\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.653294 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jzdq\" (UniqueName: \"kubernetes.io/projected/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-kube-api-access-8jzdq\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.653399 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c114d60-0f48-48f1-ba2f-e9c3e070486f-config\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.653432 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w4hz\" (UniqueName: \"kubernetes.io/projected/9c114d60-0f48-48f1-ba2f-e9c3e070486f-kube-api-access-9w4hz\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.653465 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-proxy-ca-bundles\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.653571 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c114d60-0f48-48f1-ba2f-e9c3e070486f-serving-cert\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.735905 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae1d9aa-857d-416d-92a1-e08e45bd7b19" path="/var/lib/kubelet/pods/5ae1d9aa-857d-416d-92a1-e08e45bd7b19/volumes" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.736585 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9100571-6717-43fe-a0d8-0da2bd0ce31c" path="/var/lib/kubelet/pods/a9100571-6717-43fe-a0d8-0da2bd0ce31c/volumes" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.754579 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c114d60-0f48-48f1-ba2f-e9c3e070486f-client-ca\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.754620 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jzdq\" (UniqueName: \"kubernetes.io/projected/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-kube-api-access-8jzdq\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.754649 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c114d60-0f48-48f1-ba2f-e9c3e070486f-config\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.754679 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w4hz\" (UniqueName: \"kubernetes.io/projected/9c114d60-0f48-48f1-ba2f-e9c3e070486f-kube-api-access-9w4hz\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.754713 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-proxy-ca-bundles\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.754734 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c114d60-0f48-48f1-ba2f-e9c3e070486f-serving-cert\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.754762 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-client-ca\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.754800 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-serving-cert\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.754823 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-config\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.756381 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-proxy-ca-bundles\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.756407 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c114d60-0f48-48f1-ba2f-e9c3e070486f-config\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.756425 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-client-ca\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.756419 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-config\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.756577 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c114d60-0f48-48f1-ba2f-e9c3e070486f-client-ca\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.760536 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-serving-cert\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.761106 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c114d60-0f48-48f1-ba2f-e9c3e070486f-serving-cert\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.773533 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jzdq\" (UniqueName: \"kubernetes.io/projected/0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63-kube-api-access-8jzdq\") pod \"controller-manager-75c6bd66fd-2x7bd\" (UID: \"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63\") " pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.775912 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w4hz\" (UniqueName: \"kubernetes.io/projected/9c114d60-0f48-48f1-ba2f-e9c3e070486f-kube-api-access-9w4hz\") pod \"route-controller-manager-f95c44d9b-jjf6v\" (UID: \"9c114d60-0f48-48f1-ba2f-e9c3e070486f\") " pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.924696 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:30 crc kubenswrapper[4789]: I1008 14:13:30.935344 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:31 crc kubenswrapper[4789]: I1008 14:13:31.404177 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v"] Oct 08 14:13:31 crc kubenswrapper[4789]: I1008 14:13:31.412095 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd"] Oct 08 14:13:31 crc kubenswrapper[4789]: W1008 14:13:31.414237 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f9c8fe5_6fd0_413b_8c9a_c7ad864c1e63.slice/crio-dd8d6b2671a283f500af1650f328a8011d2f934af4d55e85b71cccafb43efc39 WatchSource:0}: Error finding container dd8d6b2671a283f500af1650f328a8011d2f934af4d55e85b71cccafb43efc39: Status 404 returned error can't find the container with id dd8d6b2671a283f500af1650f328a8011d2f934af4d55e85b71cccafb43efc39 Oct 08 14:13:31 crc kubenswrapper[4789]: W1008 14:13:31.415106 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c114d60_0f48_48f1_ba2f_e9c3e070486f.slice/crio-1f0cbdca4b7aef14b1a6cafd3cc895ae70674af51c7be5b50deca28cbd0bd6b6 WatchSource:0}: Error finding container 1f0cbdca4b7aef14b1a6cafd3cc895ae70674af51c7be5b50deca28cbd0bd6b6: Status 404 returned error can't find the container with id 1f0cbdca4b7aef14b1a6cafd3cc895ae70674af51c7be5b50deca28cbd0bd6b6 Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.237978 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" event={"ID":"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63","Type":"ContainerStarted","Data":"30801d9ce3112e4d0623a1dbe3412a7ed0ada942b9c85c5b5f37b3b27779f52f"} Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.238301 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" event={"ID":"0f9c8fe5-6fd0-413b-8c9a-c7ad864c1e63","Type":"ContainerStarted","Data":"dd8d6b2671a283f500af1650f328a8011d2f934af4d55e85b71cccafb43efc39"} Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.238611 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.240770 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" event={"ID":"9c114d60-0f48-48f1-ba2f-e9c3e070486f","Type":"ContainerStarted","Data":"6e7ee0d2202a2d73800e5da878c9d6894bec7bec1f462d33996f53d9484feb89"} Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.240805 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" event={"ID":"9c114d60-0f48-48f1-ba2f-e9c3e070486f","Type":"ContainerStarted","Data":"1f0cbdca4b7aef14b1a6cafd3cc895ae70674af51c7be5b50deca28cbd0bd6b6"} Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.240972 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.242719 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.258442 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75c6bd66fd-2x7bd" podStartSLOduration=4.25842325 podStartE2EDuration="4.25842325s" podCreationTimestamp="2025-10-08 14:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:13:32.255037436 +0000 UTC m=+752.161784948" watchObservedRunningTime="2025-10-08 14:13:32.25842325 +0000 UTC m=+752.165170742" Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.286395 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" podStartSLOduration=4.286376703 podStartE2EDuration="4.286376703s" podCreationTimestamp="2025-10-08 14:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:13:32.285508579 +0000 UTC m=+752.192256071" watchObservedRunningTime="2025-10-08 14:13:32.286376703 +0000 UTC m=+752.193124195" Oct 08 14:13:32 crc kubenswrapper[4789]: I1008 14:13:32.413157 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-f95c44d9b-jjf6v" Oct 08 14:13:35 crc kubenswrapper[4789]: I1008 14:13:35.805226 4789 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 14:13:38 crc kubenswrapper[4789]: I1008 14:13:38.270035 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" event={"ID":"7141b3e4-ea4a-4476-84d2-57670fe5e932","Type":"ContainerStarted","Data":"1739b8ab7366ccc28079489bf5873fa711b9b084732120f280ecf162a53445f1"} Oct 08 14:13:38 crc kubenswrapper[4789]: I1008 14:13:38.286998 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh" podStartSLOduration=1.678113709 podStartE2EDuration="30.286962945s" podCreationTimestamp="2025-10-08 14:13:08 +0000 UTC" firstStartedPulling="2025-10-08 14:13:08.925713613 +0000 UTC m=+728.832461105" lastFinishedPulling="2025-10-08 14:13:37.534562849 +0000 UTC m=+757.441310341" observedRunningTime="2025-10-08 14:13:38.285581176 +0000 UTC m=+758.192328678" watchObservedRunningTime="2025-10-08 14:13:38.286962945 +0000 UTC m=+758.193710447" Oct 08 14:13:40 crc kubenswrapper[4789]: I1008 14:13:40.283716 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" event={"ID":"1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5","Type":"ContainerStarted","Data":"9500896b974f42e15e8804191c0a94c5fd7dcc3ebd6c0330fe1e3b06f1e90dde"} Oct 08 14:13:40 crc kubenswrapper[4789]: I1008 14:13:40.284195 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:40 crc kubenswrapper[4789]: I1008 14:13:40.299945 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" podStartSLOduration=1.9901190130000002 podStartE2EDuration="32.299927571s" podCreationTimestamp="2025-10-08 14:13:08 +0000 UTC" firstStartedPulling="2025-10-08 14:13:09.270513153 +0000 UTC m=+729.177260645" lastFinishedPulling="2025-10-08 14:13:39.580321711 +0000 UTC m=+759.487069203" observedRunningTime="2025-10-08 14:13:40.298297156 +0000 UTC m=+760.205044648" watchObservedRunningTime="2025-10-08 14:13:40.299927571 +0000 UTC m=+760.206675063" Oct 08 14:13:41 crc kubenswrapper[4789]: I1008 14:13:41.293091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" event={"ID":"9451e6d9-fa5b-4209-a332-a65412994d55","Type":"ContainerStarted","Data":"50e2fa6ea77357c2de91583236c85791d52894034ad08f61991f1587b734ae9b"} Oct 08 14:13:41 crc kubenswrapper[4789]: I1008 14:13:41.315301 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l" podStartSLOduration=-9223372003.5395 podStartE2EDuration="33.315277154s" podCreationTimestamp="2025-10-08 14:13:08 +0000 UTC" firstStartedPulling="2025-10-08 14:13:09.231152555 +0000 UTC m=+729.137900047" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:13:41.310487242 +0000 UTC m=+761.217234734" watchObservedRunningTime="2025-10-08 14:13:41.315277154 +0000 UTC m=+761.222024646" Oct 08 14:13:42 crc kubenswrapper[4789]: I1008 14:13:42.300837 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" event={"ID":"5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec","Type":"ContainerStarted","Data":"2c4903198e28286006e1d342f3b5894b1ed6ad722e419a7b153bb0d55d563a97"} Oct 08 14:13:42 crc kubenswrapper[4789]: I1008 14:13:42.302340 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" event={"ID":"643a32e3-75dd-44f5-bbac-e2d917641a29","Type":"ContainerStarted","Data":"4f74da7a0e99a310aa30ebc44bcd069ccdb6d44cc1d3868bfbb8ad20bf209f54"} Oct 08 14:13:42 crc kubenswrapper[4789]: I1008 14:13:42.302610 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:42 crc kubenswrapper[4789]: I1008 14:13:42.321431 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-lwgzt" podStartSLOduration=1.891443655 podStartE2EDuration="34.321406403s" podCreationTimestamp="2025-10-08 14:13:08 +0000 UTC" firstStartedPulling="2025-10-08 14:13:09.134193335 +0000 UTC m=+729.040940827" lastFinishedPulling="2025-10-08 14:13:41.564156083 +0000 UTC m=+761.470903575" observedRunningTime="2025-10-08 14:13:42.317897656 +0000 UTC m=+762.224645148" watchObservedRunningTime="2025-10-08 14:13:42.321406403 +0000 UTC m=+762.228153935" Oct 08 14:13:42 crc kubenswrapper[4789]: I1008 14:13:42.340204 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" podStartSLOduration=1.902209573 podStartE2EDuration="34.340185492s" podCreationTimestamp="2025-10-08 14:13:08 +0000 UTC" firstStartedPulling="2025-10-08 14:13:09.125857525 +0000 UTC m=+729.032605017" lastFinishedPulling="2025-10-08 14:13:41.563833444 +0000 UTC m=+761.470580936" observedRunningTime="2025-10-08 14:13:42.33903746 +0000 UTC m=+762.245784992" watchObservedRunningTime="2025-10-08 14:13:42.340185492 +0000 UTC m=+762.246932984" Oct 08 14:13:42 crc kubenswrapper[4789]: I1008 14:13:42.375452 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-tclgl" Oct 08 14:13:48 crc kubenswrapper[4789]: I1008 14:13:48.920748 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-rbwl7" Oct 08 14:13:56 crc kubenswrapper[4789]: I1008 14:13:56.433666 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:13:56 crc kubenswrapper[4789]: I1008 14:13:56.434061 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:13:56 crc kubenswrapper[4789]: I1008 14:13:56.434117 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:13:56 crc kubenswrapper[4789]: I1008 14:13:56.434797 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8158c6bcd8d574426927fe18b85723afd7f1a9d1d3d0dd0ab1d58a6c4269820d"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:13:56 crc kubenswrapper[4789]: I1008 14:13:56.434868 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://8158c6bcd8d574426927fe18b85723afd7f1a9d1d3d0dd0ab1d58a6c4269820d" gracePeriod=600 Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.400837 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="8158c6bcd8d574426927fe18b85723afd7f1a9d1d3d0dd0ab1d58a6c4269820d" exitCode=0 Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.400967 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"8158c6bcd8d574426927fe18b85723afd7f1a9d1d3d0dd0ab1d58a6c4269820d"} Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.401504 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"49beaccc4d4e056fc391f28224570c17529721b7b590ae8c4d56a12f0b6f40ae"} Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.401530 4789 scope.go:117] "RemoveContainer" containerID="dc34d79cea1ff807429be5448bf01d23f153ad3a6633f32da4bad1ac75061617" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.722963 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vs6x2"] Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.723976 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.736538 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-utilities\") pod \"certified-operators-vs6x2\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.736640 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-catalog-content\") pod \"certified-operators-vs6x2\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.736684 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjk2j\" (UniqueName: \"kubernetes.io/projected/08874d3a-035b-4605-8251-13b4dddc1953-kube-api-access-fjk2j\") pod \"certified-operators-vs6x2\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.739882 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vs6x2"] Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.837573 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-utilities\") pod \"certified-operators-vs6x2\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.838053 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-catalog-content\") pod \"certified-operators-vs6x2\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.838182 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-utilities\") pod \"certified-operators-vs6x2\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.838456 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-catalog-content\") pod \"certified-operators-vs6x2\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.838521 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjk2j\" (UniqueName: \"kubernetes.io/projected/08874d3a-035b-4605-8251-13b4dddc1953-kube-api-access-fjk2j\") pod \"certified-operators-vs6x2\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:57 crc kubenswrapper[4789]: I1008 14:13:57.860296 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjk2j\" (UniqueName: \"kubernetes.io/projected/08874d3a-035b-4605-8251-13b4dddc1953-kube-api-access-fjk2j\") pod \"certified-operators-vs6x2\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:58 crc kubenswrapper[4789]: I1008 14:13:58.044785 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:13:58 crc kubenswrapper[4789]: I1008 14:13:58.602713 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vs6x2"] Oct 08 14:13:58 crc kubenswrapper[4789]: W1008 14:13:58.606795 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08874d3a_035b_4605_8251_13b4dddc1953.slice/crio-2897515cbfcf7349273259109b99eb31bdca632a2a7251136ba2bffe8d75de43 WatchSource:0}: Error finding container 2897515cbfcf7349273259109b99eb31bdca632a2a7251136ba2bffe8d75de43: Status 404 returned error can't find the container with id 2897515cbfcf7349273259109b99eb31bdca632a2a7251136ba2bffe8d75de43 Oct 08 14:13:59 crc kubenswrapper[4789]: I1008 14:13:59.414779 4789 generic.go:334] "Generic (PLEG): container finished" podID="08874d3a-035b-4605-8251-13b4dddc1953" containerID="f3e020da59405436a0f3d63d89a3ea41e5db79a365c74a7154d9d7ca41442840" exitCode=0 Oct 08 14:13:59 crc kubenswrapper[4789]: I1008 14:13:59.414905 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vs6x2" event={"ID":"08874d3a-035b-4605-8251-13b4dddc1953","Type":"ContainerDied","Data":"f3e020da59405436a0f3d63d89a3ea41e5db79a365c74a7154d9d7ca41442840"} Oct 08 14:13:59 crc kubenswrapper[4789]: I1008 14:13:59.415432 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vs6x2" event={"ID":"08874d3a-035b-4605-8251-13b4dddc1953","Type":"ContainerStarted","Data":"2897515cbfcf7349273259109b99eb31bdca632a2a7251136ba2bffe8d75de43"} Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.648695 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mwbwz"] Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.650243 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.664126 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwbwz"] Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.809066 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-catalog-content\") pod \"redhat-marketplace-mwbwz\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.809103 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-utilities\") pod \"redhat-marketplace-mwbwz\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.809146 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bncg2\" (UniqueName: \"kubernetes.io/projected/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-kube-api-access-bncg2\") pod \"redhat-marketplace-mwbwz\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.910578 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-catalog-content\") pod \"redhat-marketplace-mwbwz\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.910702 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-utilities\") pod \"redhat-marketplace-mwbwz\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.910753 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bncg2\" (UniqueName: \"kubernetes.io/projected/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-kube-api-access-bncg2\") pod \"redhat-marketplace-mwbwz\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.911163 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-catalog-content\") pod \"redhat-marketplace-mwbwz\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.911163 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-utilities\") pod \"redhat-marketplace-mwbwz\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:03 crc kubenswrapper[4789]: I1008 14:14:03.929178 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bncg2\" (UniqueName: \"kubernetes.io/projected/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-kube-api-access-bncg2\") pod \"redhat-marketplace-mwbwz\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:04 crc kubenswrapper[4789]: I1008 14:14:04.008243 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:04 crc kubenswrapper[4789]: I1008 14:14:04.405036 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwbwz"] Oct 08 14:14:04 crc kubenswrapper[4789]: I1008 14:14:04.440726 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwbwz" event={"ID":"cb0048f6-1fc2-42ab-8c44-79ccc57660bc","Type":"ContainerStarted","Data":"f424e33b9059305b2345d471518c33d230d583b610424a8dc4aadca584950d11"} Oct 08 14:14:05 crc kubenswrapper[4789]: I1008 14:14:05.449855 4789 generic.go:334] "Generic (PLEG): container finished" podID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerID="647080010f90b35bfdb7cd7f093a889fa3007c1487dff7be62e6d255f4cbb0ac" exitCode=0 Oct 08 14:14:05 crc kubenswrapper[4789]: I1008 14:14:05.449933 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwbwz" event={"ID":"cb0048f6-1fc2-42ab-8c44-79ccc57660bc","Type":"ContainerDied","Data":"647080010f90b35bfdb7cd7f093a889fa3007c1487dff7be62e6d255f4cbb0ac"} Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.457533 4789 generic.go:334] "Generic (PLEG): container finished" podID="08874d3a-035b-4605-8251-13b4dddc1953" containerID="8136ab5aff0e5a8ccd07a6bdd7e7fe31f4dd77959188f67ee3bf69b771d31e64" exitCode=0 Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.457624 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vs6x2" event={"ID":"08874d3a-035b-4605-8251-13b4dddc1953","Type":"ContainerDied","Data":"8136ab5aff0e5a8ccd07a6bdd7e7fe31f4dd77959188f67ee3bf69b771d31e64"} Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.480651 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc"] Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.481816 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.484341 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.493258 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc"] Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.540505 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzw5q\" (UniqueName: \"kubernetes.io/projected/b98286a8-216e-4507-9353-f983186c2e7b-kube-api-access-lzw5q\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.540567 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.540641 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.641211 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.641279 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzw5q\" (UniqueName: \"kubernetes.io/projected/b98286a8-216e-4507-9353-f983186c2e7b-kube-api-access-lzw5q\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.641303 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.641867 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.641895 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.658632 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzw5q\" (UniqueName: \"kubernetes.io/projected/b98286a8-216e-4507-9353-f983186c2e7b-kube-api-access-lzw5q\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:06 crc kubenswrapper[4789]: I1008 14:14:06.818706 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:07 crc kubenswrapper[4789]: I1008 14:14:07.196092 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc"] Oct 08 14:14:07 crc kubenswrapper[4789]: W1008 14:14:07.202787 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb98286a8_216e_4507_9353_f983186c2e7b.slice/crio-65a7646bba776e8dd44a0e97d290b07f0f6c1f124ac44506d13c2e969e4d7f9c WatchSource:0}: Error finding container 65a7646bba776e8dd44a0e97d290b07f0f6c1f124ac44506d13c2e969e4d7f9c: Status 404 returned error can't find the container with id 65a7646bba776e8dd44a0e97d290b07f0f6c1f124ac44506d13c2e969e4d7f9c Oct 08 14:14:07 crc kubenswrapper[4789]: I1008 14:14:07.465154 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" event={"ID":"b98286a8-216e-4507-9353-f983186c2e7b","Type":"ContainerStarted","Data":"875ed695a40de82852f1b767d675a1eb23f7a082905aa8b7b90611847c1434ad"} Oct 08 14:14:07 crc kubenswrapper[4789]: I1008 14:14:07.465448 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" event={"ID":"b98286a8-216e-4507-9353-f983186c2e7b","Type":"ContainerStarted","Data":"65a7646bba776e8dd44a0e97d290b07f0f6c1f124ac44506d13c2e969e4d7f9c"} Oct 08 14:14:08 crc kubenswrapper[4789]: I1008 14:14:08.475658 4789 generic.go:334] "Generic (PLEG): container finished" podID="b98286a8-216e-4507-9353-f983186c2e7b" containerID="875ed695a40de82852f1b767d675a1eb23f7a082905aa8b7b90611847c1434ad" exitCode=0 Oct 08 14:14:08 crc kubenswrapper[4789]: I1008 14:14:08.475695 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" event={"ID":"b98286a8-216e-4507-9353-f983186c2e7b","Type":"ContainerDied","Data":"875ed695a40de82852f1b767d675a1eb23f7a082905aa8b7b90611847c1434ad"} Oct 08 14:14:09 crc kubenswrapper[4789]: I1008 14:14:09.483487 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vs6x2" event={"ID":"08874d3a-035b-4605-8251-13b4dddc1953","Type":"ContainerStarted","Data":"a48c1fd17b4d92f5dbdc76c90c5f8b0e6e0bdeefabe0c3bc914d597d4911aa33"} Oct 08 14:14:09 crc kubenswrapper[4789]: I1008 14:14:09.485413 4789 generic.go:334] "Generic (PLEG): container finished" podID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerID="d6d647c66c9513bc03b79a6ee7e170e469059d090f33734ed50976640ef7a6fa" exitCode=0 Oct 08 14:14:09 crc kubenswrapper[4789]: I1008 14:14:09.485467 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwbwz" event={"ID":"cb0048f6-1fc2-42ab-8c44-79ccc57660bc","Type":"ContainerDied","Data":"d6d647c66c9513bc03b79a6ee7e170e469059d090f33734ed50976640ef7a6fa"} Oct 08 14:14:09 crc kubenswrapper[4789]: I1008 14:14:09.504109 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vs6x2" podStartSLOduration=3.679505666 podStartE2EDuration="12.504086399s" podCreationTimestamp="2025-10-08 14:13:57 +0000 UTC" firstStartedPulling="2025-10-08 14:13:59.417771502 +0000 UTC m=+779.324518994" lastFinishedPulling="2025-10-08 14:14:08.242352235 +0000 UTC m=+788.149099727" observedRunningTime="2025-10-08 14:14:09.500409197 +0000 UTC m=+789.407156709" watchObservedRunningTime="2025-10-08 14:14:09.504086399 +0000 UTC m=+789.410833891" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.042255 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qstbz"] Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.044343 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.066738 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qstbz"] Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.087086 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-utilities\") pod \"redhat-operators-qstbz\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.087148 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbmxq\" (UniqueName: \"kubernetes.io/projected/5a3f4299-af5f-4250-9f71-0254758db380-kube-api-access-xbmxq\") pod \"redhat-operators-qstbz\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.087178 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-catalog-content\") pod \"redhat-operators-qstbz\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.188769 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-utilities\") pod \"redhat-operators-qstbz\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.188838 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbmxq\" (UniqueName: \"kubernetes.io/projected/5a3f4299-af5f-4250-9f71-0254758db380-kube-api-access-xbmxq\") pod \"redhat-operators-qstbz\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.188855 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-catalog-content\") pod \"redhat-operators-qstbz\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.189240 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-utilities\") pod \"redhat-operators-qstbz\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.189289 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-catalog-content\") pod \"redhat-operators-qstbz\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.214778 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbmxq\" (UniqueName: \"kubernetes.io/projected/5a3f4299-af5f-4250-9f71-0254758db380-kube-api-access-xbmxq\") pod \"redhat-operators-qstbz\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.403631 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.498040 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwbwz" event={"ID":"cb0048f6-1fc2-42ab-8c44-79ccc57660bc","Type":"ContainerStarted","Data":"d3ec6d1ea66f7721f3300c12fcc645b87a61071b73fe2d47ba9ca706d711d8af"} Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.526451 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mwbwz" podStartSLOduration=3.326805602 podStartE2EDuration="7.526431936s" podCreationTimestamp="2025-10-08 14:14:03 +0000 UTC" firstStartedPulling="2025-10-08 14:14:05.809281428 +0000 UTC m=+785.716028920" lastFinishedPulling="2025-10-08 14:14:10.008907762 +0000 UTC m=+789.915655254" observedRunningTime="2025-10-08 14:14:10.523178446 +0000 UTC m=+790.429925948" watchObservedRunningTime="2025-10-08 14:14:10.526431936 +0000 UTC m=+790.433179428" Oct 08 14:14:10 crc kubenswrapper[4789]: I1008 14:14:10.873236 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qstbz"] Oct 08 14:14:10 crc kubenswrapper[4789]: W1008 14:14:10.879402 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a3f4299_af5f_4250_9f71_0254758db380.slice/crio-06cdb38dbef9a1c89ee93444eba00a5ec1b62230a783b4b8e27e6204ceb75e1a WatchSource:0}: Error finding container 06cdb38dbef9a1c89ee93444eba00a5ec1b62230a783b4b8e27e6204ceb75e1a: Status 404 returned error can't find the container with id 06cdb38dbef9a1c89ee93444eba00a5ec1b62230a783b4b8e27e6204ceb75e1a Oct 08 14:14:11 crc kubenswrapper[4789]: I1008 14:14:11.504472 4789 generic.go:334] "Generic (PLEG): container finished" podID="5a3f4299-af5f-4250-9f71-0254758db380" containerID="1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9" exitCode=0 Oct 08 14:14:11 crc kubenswrapper[4789]: I1008 14:14:11.504574 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qstbz" event={"ID":"5a3f4299-af5f-4250-9f71-0254758db380","Type":"ContainerDied","Data":"1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9"} Oct 08 14:14:11 crc kubenswrapper[4789]: I1008 14:14:11.504808 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qstbz" event={"ID":"5a3f4299-af5f-4250-9f71-0254758db380","Type":"ContainerStarted","Data":"06cdb38dbef9a1c89ee93444eba00a5ec1b62230a783b4b8e27e6204ceb75e1a"} Oct 08 14:14:11 crc kubenswrapper[4789]: I1008 14:14:11.507823 4789 generic.go:334] "Generic (PLEG): container finished" podID="b98286a8-216e-4507-9353-f983186c2e7b" containerID="47bd537afc81bbcaf6a9e386f4575457f76073293e4fa18dc39f69100454b72e" exitCode=0 Oct 08 14:14:11 crc kubenswrapper[4789]: I1008 14:14:11.508303 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" event={"ID":"b98286a8-216e-4507-9353-f983186c2e7b","Type":"ContainerDied","Data":"47bd537afc81bbcaf6a9e386f4575457f76073293e4fa18dc39f69100454b72e"} Oct 08 14:14:12 crc kubenswrapper[4789]: I1008 14:14:12.515218 4789 generic.go:334] "Generic (PLEG): container finished" podID="b98286a8-216e-4507-9353-f983186c2e7b" containerID="ce9c985d1ced7763da88e6e635df03349a3af2fa0ef41e4a3f656b382de1e347" exitCode=0 Oct 08 14:14:12 crc kubenswrapper[4789]: I1008 14:14:12.515317 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" event={"ID":"b98286a8-216e-4507-9353-f983186c2e7b","Type":"ContainerDied","Data":"ce9c985d1ced7763da88e6e635df03349a3af2fa0ef41e4a3f656b382de1e347"} Oct 08 14:14:12 crc kubenswrapper[4789]: I1008 14:14:12.517739 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qstbz" event={"ID":"5a3f4299-af5f-4250-9f71-0254758db380","Type":"ContainerStarted","Data":"8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486"} Oct 08 14:14:13 crc kubenswrapper[4789]: I1008 14:14:13.525733 4789 generic.go:334] "Generic (PLEG): container finished" podID="5a3f4299-af5f-4250-9f71-0254758db380" containerID="8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486" exitCode=0 Oct 08 14:14:13 crc kubenswrapper[4789]: I1008 14:14:13.525798 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qstbz" event={"ID":"5a3f4299-af5f-4250-9f71-0254758db380","Type":"ContainerDied","Data":"8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486"} Oct 08 14:14:13 crc kubenswrapper[4789]: I1008 14:14:13.849207 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.009090 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.009414 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.033553 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-bundle\") pod \"b98286a8-216e-4507-9353-f983186c2e7b\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.033627 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-util\") pod \"b98286a8-216e-4507-9353-f983186c2e7b\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.033654 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzw5q\" (UniqueName: \"kubernetes.io/projected/b98286a8-216e-4507-9353-f983186c2e7b-kube-api-access-lzw5q\") pod \"b98286a8-216e-4507-9353-f983186c2e7b\" (UID: \"b98286a8-216e-4507-9353-f983186c2e7b\") " Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.034360 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-bundle" (OuterVolumeSpecName: "bundle") pod "b98286a8-216e-4507-9353-f983186c2e7b" (UID: "b98286a8-216e-4507-9353-f983186c2e7b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.039745 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b98286a8-216e-4507-9353-f983186c2e7b-kube-api-access-lzw5q" (OuterVolumeSpecName: "kube-api-access-lzw5q") pod "b98286a8-216e-4507-9353-f983186c2e7b" (UID: "b98286a8-216e-4507-9353-f983186c2e7b"). InnerVolumeSpecName "kube-api-access-lzw5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.044494 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-util" (OuterVolumeSpecName: "util") pod "b98286a8-216e-4507-9353-f983186c2e7b" (UID: "b98286a8-216e-4507-9353-f983186c2e7b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.047578 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.135923 4789 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.135969 4789 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b98286a8-216e-4507-9353-f983186c2e7b-util\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.135982 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzw5q\" (UniqueName: \"kubernetes.io/projected/b98286a8-216e-4507-9353-f983186c2e7b-kube-api-access-lzw5q\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.535847 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" event={"ID":"b98286a8-216e-4507-9353-f983186c2e7b","Type":"ContainerDied","Data":"65a7646bba776e8dd44a0e97d290b07f0f6c1f124ac44506d13c2e969e4d7f9c"} Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.535908 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65a7646bba776e8dd44a0e97d290b07f0f6c1f124ac44506d13c2e969e4d7f9c" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.535927 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc" Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.538769 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qstbz" event={"ID":"5a3f4299-af5f-4250-9f71-0254758db380","Type":"ContainerStarted","Data":"1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa"} Oct 08 14:14:14 crc kubenswrapper[4789]: I1008 14:14:14.568208 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qstbz" podStartSLOduration=1.8335971340000001 podStartE2EDuration="4.568182277s" podCreationTimestamp="2025-10-08 14:14:10 +0000 UTC" firstStartedPulling="2025-10-08 14:14:11.506300438 +0000 UTC m=+791.413047930" lastFinishedPulling="2025-10-08 14:14:14.240885581 +0000 UTC m=+794.147633073" observedRunningTime="2025-10-08 14:14:14.562446298 +0000 UTC m=+794.469193810" watchObservedRunningTime="2025-10-08 14:14:14.568182277 +0000 UTC m=+794.474929789" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.594863 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.952641 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2"] Oct 08 14:14:15 crc kubenswrapper[4789]: E1008 14:14:15.952905 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98286a8-216e-4507-9353-f983186c2e7b" containerName="extract" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.952925 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98286a8-216e-4507-9353-f983186c2e7b" containerName="extract" Oct 08 14:14:15 crc kubenswrapper[4789]: E1008 14:14:15.952941 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98286a8-216e-4507-9353-f983186c2e7b" containerName="pull" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.952949 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98286a8-216e-4507-9353-f983186c2e7b" containerName="pull" Oct 08 14:14:15 crc kubenswrapper[4789]: E1008 14:14:15.952966 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98286a8-216e-4507-9353-f983186c2e7b" containerName="util" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.952974 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98286a8-216e-4507-9353-f983186c2e7b" containerName="util" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.953109 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b98286a8-216e-4507-9353-f983186c2e7b" containerName="extract" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.953604 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.955623 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.955795 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.956081 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-lds6g" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.959178 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prffg\" (UniqueName: \"kubernetes.io/projected/333279f2-ccf6-44df-80ad-9097f320a3ac-kube-api-access-prffg\") pod \"nmstate-operator-858ddd8f98-cbqs2\" (UID: \"333279f2-ccf6-44df-80ad-9097f320a3ac\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2" Oct 08 14:14:15 crc kubenswrapper[4789]: I1008 14:14:15.964643 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2"] Oct 08 14:14:16 crc kubenswrapper[4789]: I1008 14:14:16.059740 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prffg\" (UniqueName: \"kubernetes.io/projected/333279f2-ccf6-44df-80ad-9097f320a3ac-kube-api-access-prffg\") pod \"nmstate-operator-858ddd8f98-cbqs2\" (UID: \"333279f2-ccf6-44df-80ad-9097f320a3ac\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2" Oct 08 14:14:16 crc kubenswrapper[4789]: I1008 14:14:16.088030 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prffg\" (UniqueName: \"kubernetes.io/projected/333279f2-ccf6-44df-80ad-9097f320a3ac-kube-api-access-prffg\") pod \"nmstate-operator-858ddd8f98-cbqs2\" (UID: \"333279f2-ccf6-44df-80ad-9097f320a3ac\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2" Oct 08 14:14:16 crc kubenswrapper[4789]: I1008 14:14:16.272063 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2" Oct 08 14:14:16 crc kubenswrapper[4789]: I1008 14:14:16.718695 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2"] Oct 08 14:14:16 crc kubenswrapper[4789]: W1008 14:14:16.735392 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod333279f2_ccf6_44df_80ad_9097f320a3ac.slice/crio-31f2eab6754ae130f1b734ed6f64461a6fc9821bf0e0284f08c99ee215a61c5e WatchSource:0}: Error finding container 31f2eab6754ae130f1b734ed6f64461a6fc9821bf0e0284f08c99ee215a61c5e: Status 404 returned error can't find the container with id 31f2eab6754ae130f1b734ed6f64461a6fc9821bf0e0284f08c99ee215a61c5e Oct 08 14:14:17 crc kubenswrapper[4789]: I1008 14:14:17.564828 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2" event={"ID":"333279f2-ccf6-44df-80ad-9097f320a3ac","Type":"ContainerStarted","Data":"31f2eab6754ae130f1b734ed6f64461a6fc9821bf0e0284f08c99ee215a61c5e"} Oct 08 14:14:18 crc kubenswrapper[4789]: I1008 14:14:18.045890 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:14:18 crc kubenswrapper[4789]: I1008 14:14:18.045966 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:14:18 crc kubenswrapper[4789]: I1008 14:14:18.093942 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:14:18 crc kubenswrapper[4789]: I1008 14:14:18.434570 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwbwz"] Oct 08 14:14:18 crc kubenswrapper[4789]: I1008 14:14:18.434822 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mwbwz" podUID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerName="registry-server" containerID="cri-o://d3ec6d1ea66f7721f3300c12fcc645b87a61071b73fe2d47ba9ca706d711d8af" gracePeriod=2 Oct 08 14:14:18 crc kubenswrapper[4789]: I1008 14:14:18.613176 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.403939 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.404239 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.465933 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.591264 4789 generic.go:334] "Generic (PLEG): container finished" podID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerID="d3ec6d1ea66f7721f3300c12fcc645b87a61071b73fe2d47ba9ca706d711d8af" exitCode=0 Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.591447 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwbwz" event={"ID":"cb0048f6-1fc2-42ab-8c44-79ccc57660bc","Type":"ContainerDied","Data":"d3ec6d1ea66f7721f3300c12fcc645b87a61071b73fe2d47ba9ca706d711d8af"} Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.632578 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.785566 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.831230 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-catalog-content\") pod \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.831397 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-utilities\") pod \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.831454 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bncg2\" (UniqueName: \"kubernetes.io/projected/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-kube-api-access-bncg2\") pod \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\" (UID: \"cb0048f6-1fc2-42ab-8c44-79ccc57660bc\") " Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.832177 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-utilities" (OuterVolumeSpecName: "utilities") pod "cb0048f6-1fc2-42ab-8c44-79ccc57660bc" (UID: "cb0048f6-1fc2-42ab-8c44-79ccc57660bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.836257 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-kube-api-access-bncg2" (OuterVolumeSpecName: "kube-api-access-bncg2") pod "cb0048f6-1fc2-42ab-8c44-79ccc57660bc" (UID: "cb0048f6-1fc2-42ab-8c44-79ccc57660bc"). InnerVolumeSpecName "kube-api-access-bncg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.842850 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb0048f6-1fc2-42ab-8c44-79ccc57660bc" (UID: "cb0048f6-1fc2-42ab-8c44-79ccc57660bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.932792 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.932825 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bncg2\" (UniqueName: \"kubernetes.io/projected/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-kube-api-access-bncg2\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:20 crc kubenswrapper[4789]: I1008 14:14:20.932835 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb0048f6-1fc2-42ab-8c44-79ccc57660bc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:21 crc kubenswrapper[4789]: I1008 14:14:21.598014 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2" event={"ID":"333279f2-ccf6-44df-80ad-9097f320a3ac","Type":"ContainerStarted","Data":"94fa0a12e9352552744e77e35a0f5a14ea1575ace530f711f6a86affb074b969"} Oct 08 14:14:21 crc kubenswrapper[4789]: I1008 14:14:21.600464 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwbwz" Oct 08 14:14:21 crc kubenswrapper[4789]: I1008 14:14:21.603090 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwbwz" event={"ID":"cb0048f6-1fc2-42ab-8c44-79ccc57660bc","Type":"ContainerDied","Data":"f424e33b9059305b2345d471518c33d230d583b610424a8dc4aadca584950d11"} Oct 08 14:14:21 crc kubenswrapper[4789]: I1008 14:14:21.603135 4789 scope.go:117] "RemoveContainer" containerID="d3ec6d1ea66f7721f3300c12fcc645b87a61071b73fe2d47ba9ca706d711d8af" Oct 08 14:14:21 crc kubenswrapper[4789]: I1008 14:14:21.621736 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-cbqs2" podStartSLOduration=2.050896677 podStartE2EDuration="6.621718571s" podCreationTimestamp="2025-10-08 14:14:15 +0000 UTC" firstStartedPulling="2025-10-08 14:14:16.736660372 +0000 UTC m=+796.643407864" lastFinishedPulling="2025-10-08 14:14:21.307482266 +0000 UTC m=+801.214229758" observedRunningTime="2025-10-08 14:14:21.619094848 +0000 UTC m=+801.525842340" watchObservedRunningTime="2025-10-08 14:14:21.621718571 +0000 UTC m=+801.528466063" Oct 08 14:14:21 crc kubenswrapper[4789]: I1008 14:14:21.637528 4789 scope.go:117] "RemoveContainer" containerID="d6d647c66c9513bc03b79a6ee7e170e469059d090f33734ed50976640ef7a6fa" Oct 08 14:14:21 crc kubenswrapper[4789]: I1008 14:14:21.640480 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwbwz"] Oct 08 14:14:21 crc kubenswrapper[4789]: I1008 14:14:21.643533 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwbwz"] Oct 08 14:14:21 crc kubenswrapper[4789]: I1008 14:14:21.666355 4789 scope.go:117] "RemoveContainer" containerID="647080010f90b35bfdb7cd7f093a889fa3007c1487dff7be62e6d255f4cbb0ac" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.230700 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vs6x2"] Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.230972 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vs6x2" podUID="08874d3a-035b-4605-8251-13b4dddc1953" containerName="registry-server" containerID="cri-o://a48c1fd17b4d92f5dbdc76c90c5f8b0e6e0bdeefabe0c3bc914d597d4911aa33" gracePeriod=2 Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.498968 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj"] Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.499431 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerName="registry-server" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.499442 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerName="registry-server" Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.499452 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerName="extract-utilities" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.499458 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerName="extract-utilities" Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.499472 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerName="extract-content" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.499478 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerName="extract-content" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.499566 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" containerName="registry-server" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.500190 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.503792 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-ntpf2" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.526837 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj"] Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.533885 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-2qnx8"] Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.534654 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.536932 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl"] Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.540851 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.543602 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.560638 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtxb2\" (UniqueName: \"kubernetes.io/projected/e278e129-8d89-4095-a7f0-88a44cb19c97-kube-api-access-vtxb2\") pod \"nmstate-metrics-fdff9cb8d-l7cdj\" (UID: \"e278e129-8d89-4095-a7f0-88a44cb19c97\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.560753 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/764cebf9-8453-4c3b-afe4-c6f56b39d87f-ovs-socket\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.560814 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/764cebf9-8453-4c3b-afe4-c6f56b39d87f-nmstate-lock\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.560869 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqhm6\" (UniqueName: \"kubernetes.io/projected/764cebf9-8453-4c3b-afe4-c6f56b39d87f-kube-api-access-jqhm6\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.560927 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/764cebf9-8453-4c3b-afe4-c6f56b39d87f-dbus-socket\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.597857 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl"] Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.609821 4789 generic.go:334] "Generic (PLEG): container finished" podID="08874d3a-035b-4605-8251-13b4dddc1953" containerID="a48c1fd17b4d92f5dbdc76c90c5f8b0e6e0bdeefabe0c3bc914d597d4911aa33" exitCode=0 Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.609923 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vs6x2" event={"ID":"08874d3a-035b-4605-8251-13b4dddc1953","Type":"ContainerDied","Data":"a48c1fd17b4d92f5dbdc76c90c5f8b0e6e0bdeefabe0c3bc914d597d4911aa33"} Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.609950 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vs6x2" event={"ID":"08874d3a-035b-4605-8251-13b4dddc1953","Type":"ContainerDied","Data":"2897515cbfcf7349273259109b99eb31bdca632a2a7251136ba2bffe8d75de43"} Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.609960 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2897515cbfcf7349273259109b99eb31bdca632a2a7251136ba2bffe8d75de43" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.644202 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh"] Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.645384 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.649312 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.650511 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh"] Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.651562 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.651877 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.652122 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-gbc8w" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.663246 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/764cebf9-8453-4c3b-afe4-c6f56b39d87f-nmstate-lock\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.663300 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/066fb44b-7e80-4cea-b5c9-ab9e208a9d52-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-8ndgl\" (UID: \"066fb44b-7e80-4cea-b5c9-ab9e208a9d52\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.663347 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqhm6\" (UniqueName: \"kubernetes.io/projected/764cebf9-8453-4c3b-afe4-c6f56b39d87f-kube-api-access-jqhm6\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.663388 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/764cebf9-8453-4c3b-afe4-c6f56b39d87f-dbus-socket\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.663445 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtxb2\" (UniqueName: \"kubernetes.io/projected/e278e129-8d89-4095-a7f0-88a44cb19c97-kube-api-access-vtxb2\") pod \"nmstate-metrics-fdff9cb8d-l7cdj\" (UID: \"e278e129-8d89-4095-a7f0-88a44cb19c97\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.663488 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5zb2\" (UniqueName: \"kubernetes.io/projected/066fb44b-7e80-4cea-b5c9-ab9e208a9d52-kube-api-access-h5zb2\") pod \"nmstate-webhook-6cdbc54649-8ndgl\" (UID: \"066fb44b-7e80-4cea-b5c9-ab9e208a9d52\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.663536 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/764cebf9-8453-4c3b-afe4-c6f56b39d87f-ovs-socket\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.663622 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/764cebf9-8453-4c3b-afe4-c6f56b39d87f-ovs-socket\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.664140 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/764cebf9-8453-4c3b-afe4-c6f56b39d87f-nmstate-lock\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.664609 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/764cebf9-8453-4c3b-afe4-c6f56b39d87f-dbus-socket\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.757823 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb0048f6-1fc2-42ab-8c44-79ccc57660bc" path="/var/lib/kubelet/pods/cb0048f6-1fc2-42ab-8c44-79ccc57660bc/volumes" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.764664 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-utilities\") pod \"08874d3a-035b-4605-8251-13b4dddc1953\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.764742 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjk2j\" (UniqueName: \"kubernetes.io/projected/08874d3a-035b-4605-8251-13b4dddc1953-kube-api-access-fjk2j\") pod \"08874d3a-035b-4605-8251-13b4dddc1953\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.764804 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-catalog-content\") pod \"08874d3a-035b-4605-8251-13b4dddc1953\" (UID: \"08874d3a-035b-4605-8251-13b4dddc1953\") " Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.764980 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/066fb44b-7e80-4cea-b5c9-ab9e208a9d52-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-8ndgl\" (UID: \"066fb44b-7e80-4cea-b5c9-ab9e208a9d52\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.765039 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b5615e4f-312c-49dc-9689-3662e0ad7ac8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.765272 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbrkg\" (UniqueName: \"kubernetes.io/projected/b5615e4f-312c-49dc-9689-3662e0ad7ac8-kube-api-access-qbrkg\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.765336 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5615e4f-312c-49dc-9689-3662e0ad7ac8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.765446 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5zb2\" (UniqueName: \"kubernetes.io/projected/066fb44b-7e80-4cea-b5c9-ab9e208a9d52-kube-api-access-h5zb2\") pod \"nmstate-webhook-6cdbc54649-8ndgl\" (UID: \"066fb44b-7e80-4cea-b5c9-ab9e208a9d52\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.766591 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-utilities" (OuterVolumeSpecName: "utilities") pod "08874d3a-035b-4605-8251-13b4dddc1953" (UID: "08874d3a-035b-4605-8251-13b4dddc1953"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.772088 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtxb2\" (UniqueName: \"kubernetes.io/projected/e278e129-8d89-4095-a7f0-88a44cb19c97-kube-api-access-vtxb2\") pod \"nmstate-metrics-fdff9cb8d-l7cdj\" (UID: \"e278e129-8d89-4095-a7f0-88a44cb19c97\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj" Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.772476 4789 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.772536 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/066fb44b-7e80-4cea-b5c9-ab9e208a9d52-tls-key-pair podName:066fb44b-7e80-4cea-b5c9-ab9e208a9d52 nodeName:}" failed. No retries permitted until 2025-10-08 14:14:23.272519177 +0000 UTC m=+803.179266659 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/066fb44b-7e80-4cea-b5c9-ab9e208a9d52-tls-key-pair") pod "nmstate-webhook-6cdbc54649-8ndgl" (UID: "066fb44b-7e80-4cea-b5c9-ab9e208a9d52") : secret "openshift-nmstate-webhook" not found Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.773517 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08874d3a-035b-4605-8251-13b4dddc1953-kube-api-access-fjk2j" (OuterVolumeSpecName: "kube-api-access-fjk2j") pod "08874d3a-035b-4605-8251-13b4dddc1953" (UID: "08874d3a-035b-4605-8251-13b4dddc1953"). InnerVolumeSpecName "kube-api-access-fjk2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.780417 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqhm6\" (UniqueName: \"kubernetes.io/projected/764cebf9-8453-4c3b-afe4-c6f56b39d87f-kube-api-access-jqhm6\") pod \"nmstate-handler-2qnx8\" (UID: \"764cebf9-8453-4c3b-afe4-c6f56b39d87f\") " pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.789265 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5zb2\" (UniqueName: \"kubernetes.io/projected/066fb44b-7e80-4cea-b5c9-ab9e208a9d52-kube-api-access-h5zb2\") pod \"nmstate-webhook-6cdbc54649-8ndgl\" (UID: \"066fb44b-7e80-4cea-b5c9-ab9e208a9d52\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.830663 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.833104 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08874d3a-035b-4605-8251-13b4dddc1953" (UID: "08874d3a-035b-4605-8251-13b4dddc1953"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.866754 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b5615e4f-312c-49dc-9689-3662e0ad7ac8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.867153 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbrkg\" (UniqueName: \"kubernetes.io/projected/b5615e4f-312c-49dc-9689-3662e0ad7ac8-kube-api-access-qbrkg\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.867198 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5615e4f-312c-49dc-9689-3662e0ad7ac8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.867263 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.867277 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjk2j\" (UniqueName: \"kubernetes.io/projected/08874d3a-035b-4605-8251-13b4dddc1953-kube-api-access-fjk2j\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.867288 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08874d3a-035b-4605-8251-13b4dddc1953-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.867385 4789 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.867434 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5615e4f-312c-49dc-9689-3662e0ad7ac8-plugin-serving-cert podName:b5615e4f-312c-49dc-9689-3662e0ad7ac8 nodeName:}" failed. No retries permitted until 2025-10-08 14:14:23.36741866 +0000 UTC m=+803.274166152 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/b5615e4f-312c-49dc-9689-3662e0ad7ac8-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-4cxwh" (UID: "b5615e4f-312c-49dc-9689-3662e0ad7ac8") : secret "plugin-serving-cert" not found Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.868021 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b5615e4f-312c-49dc-9689-3662e0ad7ac8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.872369 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.908888 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbrkg\" (UniqueName: \"kubernetes.io/projected/b5615e4f-312c-49dc-9689-3662e0ad7ac8-kube-api-access-qbrkg\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.912447 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-97bccbb57-pcl6c"] Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.912666 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08874d3a-035b-4605-8251-13b4dddc1953" containerName="registry-server" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.912682 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="08874d3a-035b-4605-8251-13b4dddc1953" containerName="registry-server" Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.912694 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08874d3a-035b-4605-8251-13b4dddc1953" containerName="extract-utilities" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.912700 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="08874d3a-035b-4605-8251-13b4dddc1953" containerName="extract-utilities" Oct 08 14:14:22 crc kubenswrapper[4789]: E1008 14:14:22.912715 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08874d3a-035b-4605-8251-13b4dddc1953" containerName="extract-content" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.912723 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="08874d3a-035b-4605-8251-13b4dddc1953" containerName="extract-content" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.912827 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="08874d3a-035b-4605-8251-13b4dddc1953" containerName="registry-server" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.913231 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.929356 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-97bccbb57-pcl6c"] Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.968880 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-oauth-serving-cert\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.968932 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa489c05-b844-4d52-9bb2-495db2f40966-console-serving-cert\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.968969 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r527\" (UniqueName: \"kubernetes.io/projected/fa489c05-b844-4d52-9bb2-495db2f40966-kube-api-access-8r527\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.969049 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa489c05-b844-4d52-9bb2-495db2f40966-console-oauth-config\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.969105 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-console-config\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.969132 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-trusted-ca-bundle\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:22 crc kubenswrapper[4789]: I1008 14:14:22.969159 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-service-ca\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.070118 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-oauth-serving-cert\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.070164 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa489c05-b844-4d52-9bb2-495db2f40966-console-serving-cert\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.070197 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r527\" (UniqueName: \"kubernetes.io/projected/fa489c05-b844-4d52-9bb2-495db2f40966-kube-api-access-8r527\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.070251 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa489c05-b844-4d52-9bb2-495db2f40966-console-oauth-config\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.070296 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-console-config\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.070317 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-trusted-ca-bundle\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.070338 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-service-ca\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.071825 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-service-ca\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.072023 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-console-config\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.072479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-trusted-ca-bundle\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.072552 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa489c05-b844-4d52-9bb2-495db2f40966-oauth-serving-cert\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.074478 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa489c05-b844-4d52-9bb2-495db2f40966-console-serving-cert\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.076578 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa489c05-b844-4d52-9bb2-495db2f40966-console-oauth-config\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.089194 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r527\" (UniqueName: \"kubernetes.io/projected/fa489c05-b844-4d52-9bb2-495db2f40966-kube-api-access-8r527\") pod \"console-97bccbb57-pcl6c\" (UID: \"fa489c05-b844-4d52-9bb2-495db2f40966\") " pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.235327 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.335819 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj"] Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.373472 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/066fb44b-7e80-4cea-b5c9-ab9e208a9d52-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-8ndgl\" (UID: \"066fb44b-7e80-4cea-b5c9-ab9e208a9d52\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.373569 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5615e4f-312c-49dc-9689-3662e0ad7ac8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.377393 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/066fb44b-7e80-4cea-b5c9-ab9e208a9d52-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-8ndgl\" (UID: \"066fb44b-7e80-4cea-b5c9-ab9e208a9d52\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.378071 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5615e4f-312c-49dc-9689-3662e0ad7ac8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-4cxwh\" (UID: \"b5615e4f-312c-49dc-9689-3662e0ad7ac8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.435223 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-97bccbb57-pcl6c"] Oct 08 14:14:23 crc kubenswrapper[4789]: W1008 14:14:23.441776 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa489c05_b844_4d52_9bb2_495db2f40966.slice/crio-cd3c568aae31a4bca4a22fd4901bec3ab0e4e8cea44f2405695e8ae8e981f298 WatchSource:0}: Error finding container cd3c568aae31a4bca4a22fd4901bec3ab0e4e8cea44f2405695e8ae8e981f298: Status 404 returned error can't find the container with id cd3c568aae31a4bca4a22fd4901bec3ab0e4e8cea44f2405695e8ae8e981f298 Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.482550 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.577707 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.623869 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2qnx8" event={"ID":"764cebf9-8453-4c3b-afe4-c6f56b39d87f","Type":"ContainerStarted","Data":"803a3d9f520e92a509a5e44716bb510f12d172c539a1a1ed609707606eb246fd"} Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.625218 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-97bccbb57-pcl6c" event={"ID":"fa489c05-b844-4d52-9bb2-495db2f40966","Type":"ContainerStarted","Data":"7a92462e08647b1460d1cd1a36b3ef219fa68f6756aecaccc66b424f035c88fb"} Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.625236 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-97bccbb57-pcl6c" event={"ID":"fa489c05-b844-4d52-9bb2-495db2f40966","Type":"ContainerStarted","Data":"cd3c568aae31a4bca4a22fd4901bec3ab0e4e8cea44f2405695e8ae8e981f298"} Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.625909 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vs6x2" Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.628359 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj" event={"ID":"e278e129-8d89-4095-a7f0-88a44cb19c97","Type":"ContainerStarted","Data":"522feb56c2b7db4830e5f26ec2192e8802e17542bcdecbc535fc43c6ab92aa66"} Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.658655 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vs6x2"] Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.663169 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vs6x2"] Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.816249 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh"] Oct 08 14:14:23 crc kubenswrapper[4789]: I1008 14:14:23.867962 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl"] Oct 08 14:14:24 crc kubenswrapper[4789]: I1008 14:14:24.641751 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" event={"ID":"066fb44b-7e80-4cea-b5c9-ab9e208a9d52","Type":"ContainerStarted","Data":"e3c73a0ad225badde4bcf3a4c98f574940faf10b698ee5b2a0259a24f091f78c"} Oct 08 14:14:24 crc kubenswrapper[4789]: I1008 14:14:24.642952 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" event={"ID":"b5615e4f-312c-49dc-9689-3662e0ad7ac8","Type":"ContainerStarted","Data":"5b53c20413a4d34147ef7e195b835cb9d2fbb8cbd1129111dfc9995de688d1dc"} Oct 08 14:14:24 crc kubenswrapper[4789]: I1008 14:14:24.663150 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-97bccbb57-pcl6c" podStartSLOduration=2.663130733 podStartE2EDuration="2.663130733s" podCreationTimestamp="2025-10-08 14:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:14:24.660570322 +0000 UTC m=+804.567317854" watchObservedRunningTime="2025-10-08 14:14:24.663130733 +0000 UTC m=+804.569878235" Oct 08 14:14:24 crc kubenswrapper[4789]: I1008 14:14:24.745365 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08874d3a-035b-4605-8251-13b4dddc1953" path="/var/lib/kubelet/pods/08874d3a-035b-4605-8251-13b4dddc1953/volumes" Oct 08 14:14:25 crc kubenswrapper[4789]: I1008 14:14:25.832026 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qstbz"] Oct 08 14:14:25 crc kubenswrapper[4789]: I1008 14:14:25.832517 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qstbz" podUID="5a3f4299-af5f-4250-9f71-0254758db380" containerName="registry-server" containerID="cri-o://1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa" gracePeriod=2 Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.593293 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.615807 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbmxq\" (UniqueName: \"kubernetes.io/projected/5a3f4299-af5f-4250-9f71-0254758db380-kube-api-access-xbmxq\") pod \"5a3f4299-af5f-4250-9f71-0254758db380\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.615868 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-catalog-content\") pod \"5a3f4299-af5f-4250-9f71-0254758db380\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.615896 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-utilities\") pod \"5a3f4299-af5f-4250-9f71-0254758db380\" (UID: \"5a3f4299-af5f-4250-9f71-0254758db380\") " Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.617040 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-utilities" (OuterVolumeSpecName: "utilities") pod "5a3f4299-af5f-4250-9f71-0254758db380" (UID: "5a3f4299-af5f-4250-9f71-0254758db380"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.620154 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a3f4299-af5f-4250-9f71-0254758db380-kube-api-access-xbmxq" (OuterVolumeSpecName: "kube-api-access-xbmxq") pod "5a3f4299-af5f-4250-9f71-0254758db380" (UID: "5a3f4299-af5f-4250-9f71-0254758db380"). InnerVolumeSpecName "kube-api-access-xbmxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.654290 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj" event={"ID":"e278e129-8d89-4095-a7f0-88a44cb19c97","Type":"ContainerStarted","Data":"6ceca5ba17b27c2a3384471cb1ead9cae6d0014f9444b3fdff321ee49c72e435"} Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.656337 4789 generic.go:334] "Generic (PLEG): container finished" podID="5a3f4299-af5f-4250-9f71-0254758db380" containerID="1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa" exitCode=0 Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.656372 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qstbz" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.656398 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qstbz" event={"ID":"5a3f4299-af5f-4250-9f71-0254758db380","Type":"ContainerDied","Data":"1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa"} Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.656429 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qstbz" event={"ID":"5a3f4299-af5f-4250-9f71-0254758db380","Type":"ContainerDied","Data":"06cdb38dbef9a1c89ee93444eba00a5ec1b62230a783b4b8e27e6204ceb75e1a"} Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.656455 4789 scope.go:117] "RemoveContainer" containerID="1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.660496 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" event={"ID":"b5615e4f-312c-49dc-9689-3662e0ad7ac8","Type":"ContainerStarted","Data":"fc120bc8704395f89887291b918e1346d5001b9e661c5428df01bf6edcddcbf0"} Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.683056 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4cxwh" podStartSLOduration=2.063594692 podStartE2EDuration="4.683037651s" podCreationTimestamp="2025-10-08 14:14:22 +0000 UTC" firstStartedPulling="2025-10-08 14:14:23.826335634 +0000 UTC m=+803.733083126" lastFinishedPulling="2025-10-08 14:14:26.445778593 +0000 UTC m=+806.352526085" observedRunningTime="2025-10-08 14:14:26.67466384 +0000 UTC m=+806.581411332" watchObservedRunningTime="2025-10-08 14:14:26.683037651 +0000 UTC m=+806.589785143" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.718299 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbmxq\" (UniqueName: \"kubernetes.io/projected/5a3f4299-af5f-4250-9f71-0254758db380-kube-api-access-xbmxq\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.718350 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.718740 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a3f4299-af5f-4250-9f71-0254758db380" (UID: "5a3f4299-af5f-4250-9f71-0254758db380"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.819812 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3f4299-af5f-4250-9f71-0254758db380-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.976206 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qstbz"] Oct 08 14:14:26 crc kubenswrapper[4789]: I1008 14:14:26.981441 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qstbz"] Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.282479 4789 scope.go:117] "RemoveContainer" containerID="8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.307522 4789 scope.go:117] "RemoveContainer" containerID="1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.352306 4789 scope.go:117] "RemoveContainer" containerID="1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa" Oct 08 14:14:27 crc kubenswrapper[4789]: E1008 14:14:27.353079 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa\": container with ID starting with 1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa not found: ID does not exist" containerID="1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.353122 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa"} err="failed to get container status \"1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa\": rpc error: code = NotFound desc = could not find container \"1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa\": container with ID starting with 1dd5438a09fdc66cd5cdd8f66fa09ba1aba476fb66e41afc0582c733ebb55afa not found: ID does not exist" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.353146 4789 scope.go:117] "RemoveContainer" containerID="8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486" Oct 08 14:14:27 crc kubenswrapper[4789]: E1008 14:14:27.353458 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486\": container with ID starting with 8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486 not found: ID does not exist" containerID="8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.353492 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486"} err="failed to get container status \"8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486\": rpc error: code = NotFound desc = could not find container \"8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486\": container with ID starting with 8bff26c9cece2ee9f2d98a42095969cdb095aa6d4917991b425143091d579486 not found: ID does not exist" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.353513 4789 scope.go:117] "RemoveContainer" containerID="1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9" Oct 08 14:14:27 crc kubenswrapper[4789]: E1008 14:14:27.354821 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9\": container with ID starting with 1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9 not found: ID does not exist" containerID="1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.354849 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9"} err="failed to get container status \"1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9\": rpc error: code = NotFound desc = could not find container \"1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9\": container with ID starting with 1639525d9cb9a4ce28f51415e3c13f4b45e49025198d26c806e792f0debb67c9 not found: ID does not exist" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.668600 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" event={"ID":"066fb44b-7e80-4cea-b5c9-ab9e208a9d52","Type":"ContainerStarted","Data":"322d495b04af6e9a560f2ebfb8423a9660e21afc82235f2d966be753da066349"} Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.668886 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.672543 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2qnx8" event={"ID":"764cebf9-8453-4c3b-afe4-c6f56b39d87f","Type":"ContainerStarted","Data":"c2a23f742a8eda86501533e6fe4036255f9d468a189998a577093d35a29cc7fe"} Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.693479 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" podStartSLOduration=2.28311838 podStartE2EDuration="5.693457149s" podCreationTimestamp="2025-10-08 14:14:22 +0000 UTC" firstStartedPulling="2025-10-08 14:14:23.874230118 +0000 UTC m=+803.780977610" lastFinishedPulling="2025-10-08 14:14:27.284568887 +0000 UTC m=+807.191316379" observedRunningTime="2025-10-08 14:14:27.691083683 +0000 UTC m=+807.597831185" watchObservedRunningTime="2025-10-08 14:14:27.693457149 +0000 UTC m=+807.600204641" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.717459 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-2qnx8" podStartSLOduration=1.353045463 podStartE2EDuration="5.717442741s" podCreationTimestamp="2025-10-08 14:14:22 +0000 UTC" firstStartedPulling="2025-10-08 14:14:22.91913167 +0000 UTC m=+802.825879162" lastFinishedPulling="2025-10-08 14:14:27.283528918 +0000 UTC m=+807.190276440" observedRunningTime="2025-10-08 14:14:27.712546946 +0000 UTC m=+807.619294438" watchObservedRunningTime="2025-10-08 14:14:27.717442741 +0000 UTC m=+807.624190233" Oct 08 14:14:27 crc kubenswrapper[4789]: I1008 14:14:27.873480 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:28 crc kubenswrapper[4789]: I1008 14:14:28.740337 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a3f4299-af5f-4250-9f71-0254758db380" path="/var/lib/kubelet/pods/5a3f4299-af5f-4250-9f71-0254758db380/volumes" Oct 08 14:14:30 crc kubenswrapper[4789]: I1008 14:14:30.736272 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj" event={"ID":"e278e129-8d89-4095-a7f0-88a44cb19c97","Type":"ContainerStarted","Data":"4923088088f327ab6f9843605a0520e80a7564114a0d3b8e154182ceeb4b425f"} Oct 08 14:14:30 crc kubenswrapper[4789]: I1008 14:14:30.765603 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-l7cdj" podStartSLOduration=2.40761502 podStartE2EDuration="8.765580199s" podCreationTimestamp="2025-10-08 14:14:22 +0000 UTC" firstStartedPulling="2025-10-08 14:14:23.344322512 +0000 UTC m=+803.251070004" lastFinishedPulling="2025-10-08 14:14:29.702287691 +0000 UTC m=+809.609035183" observedRunningTime="2025-10-08 14:14:30.763255435 +0000 UTC m=+810.670002947" watchObservedRunningTime="2025-10-08 14:14:30.765580199 +0000 UTC m=+810.672327691" Oct 08 14:14:32 crc kubenswrapper[4789]: I1008 14:14:32.895644 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-2qnx8" Oct 08 14:14:33 crc kubenswrapper[4789]: I1008 14:14:33.235844 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:33 crc kubenswrapper[4789]: I1008 14:14:33.235914 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:33 crc kubenswrapper[4789]: I1008 14:14:33.241935 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:33 crc kubenswrapper[4789]: I1008 14:14:33.756662 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-97bccbb57-pcl6c" Oct 08 14:14:33 crc kubenswrapper[4789]: I1008 14:14:33.815220 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-dwjgm"] Oct 08 14:14:43 crc kubenswrapper[4789]: I1008 14:14:43.491141 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-8ndgl" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.024848 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2dgnw"] Oct 08 14:14:47 crc kubenswrapper[4789]: E1008 14:14:47.025396 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3f4299-af5f-4250-9f71-0254758db380" containerName="extract-content" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.025407 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3f4299-af5f-4250-9f71-0254758db380" containerName="extract-content" Oct 08 14:14:47 crc kubenswrapper[4789]: E1008 14:14:47.025416 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3f4299-af5f-4250-9f71-0254758db380" containerName="registry-server" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.025424 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3f4299-af5f-4250-9f71-0254758db380" containerName="registry-server" Oct 08 14:14:47 crc kubenswrapper[4789]: E1008 14:14:47.025435 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3f4299-af5f-4250-9f71-0254758db380" containerName="extract-utilities" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.025441 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3f4299-af5f-4250-9f71-0254758db380" containerName="extract-utilities" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.025544 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a3f4299-af5f-4250-9f71-0254758db380" containerName="registry-server" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.026386 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.033810 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2dgnw"] Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.200095 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-utilities\") pod \"community-operators-2dgnw\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.200157 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-catalog-content\") pod \"community-operators-2dgnw\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.200205 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldr98\" (UniqueName: \"kubernetes.io/projected/82dd900f-00ff-4626-b185-6101ea04a6fd-kube-api-access-ldr98\") pod \"community-operators-2dgnw\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.300862 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-utilities\") pod \"community-operators-2dgnw\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.300904 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-catalog-content\") pod \"community-operators-2dgnw\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.300943 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldr98\" (UniqueName: \"kubernetes.io/projected/82dd900f-00ff-4626-b185-6101ea04a6fd-kube-api-access-ldr98\") pod \"community-operators-2dgnw\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.301373 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-utilities\") pod \"community-operators-2dgnw\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.301428 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-catalog-content\") pod \"community-operators-2dgnw\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.328791 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldr98\" (UniqueName: \"kubernetes.io/projected/82dd900f-00ff-4626-b185-6101ea04a6fd-kube-api-access-ldr98\") pod \"community-operators-2dgnw\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.346675 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.826853 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2dgnw"] Oct 08 14:14:47 crc kubenswrapper[4789]: I1008 14:14:47.852338 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2dgnw" event={"ID":"82dd900f-00ff-4626-b185-6101ea04a6fd","Type":"ContainerStarted","Data":"23121469d6c525b8bc55fea67d9ed2545e3ff7eaccf1150480fbe08345841394"} Oct 08 14:14:48 crc kubenswrapper[4789]: I1008 14:14:48.860120 4789 generic.go:334] "Generic (PLEG): container finished" podID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerID="051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4" exitCode=0 Oct 08 14:14:48 crc kubenswrapper[4789]: I1008 14:14:48.860160 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2dgnw" event={"ID":"82dd900f-00ff-4626-b185-6101ea04a6fd","Type":"ContainerDied","Data":"051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4"} Oct 08 14:14:50 crc kubenswrapper[4789]: I1008 14:14:50.879095 4789 generic.go:334] "Generic (PLEG): container finished" podID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerID="b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac" exitCode=0 Oct 08 14:14:50 crc kubenswrapper[4789]: I1008 14:14:50.879477 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2dgnw" event={"ID":"82dd900f-00ff-4626-b185-6101ea04a6fd","Type":"ContainerDied","Data":"b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac"} Oct 08 14:14:51 crc kubenswrapper[4789]: I1008 14:14:51.913836 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2dgnw" event={"ID":"82dd900f-00ff-4626-b185-6101ea04a6fd","Type":"ContainerStarted","Data":"2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052"} Oct 08 14:14:51 crc kubenswrapper[4789]: I1008 14:14:51.935597 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2dgnw" podStartSLOduration=3.383875633 podStartE2EDuration="5.935582805s" podCreationTimestamp="2025-10-08 14:14:46 +0000 UTC" firstStartedPulling="2025-10-08 14:14:48.861896413 +0000 UTC m=+828.768643905" lastFinishedPulling="2025-10-08 14:14:51.413603585 +0000 UTC m=+831.320351077" observedRunningTime="2025-10-08 14:14:51.935178954 +0000 UTC m=+831.841926446" watchObservedRunningTime="2025-10-08 14:14:51.935582805 +0000 UTC m=+831.842330287" Oct 08 14:14:57 crc kubenswrapper[4789]: I1008 14:14:57.347432 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:57 crc kubenswrapper[4789]: I1008 14:14:57.347935 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:57 crc kubenswrapper[4789]: I1008 14:14:57.412947 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:57 crc kubenswrapper[4789]: I1008 14:14:57.988774 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.026455 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2dgnw"] Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.690008 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2"] Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.691514 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.692943 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.702493 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2"] Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.748002 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm59g\" (UniqueName: \"kubernetes.io/projected/5084833a-0245-4282-8816-9d722342b260-kube-api-access-fm59g\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.748357 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.748391 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.848925 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm59g\" (UniqueName: \"kubernetes.io/projected/5084833a-0245-4282-8816-9d722342b260-kube-api-access-fm59g\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.848975 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.849030 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.849607 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.849756 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.866699 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-dwjgm" podUID="6fd06952-1b3e-44e6-af19-dfa31965b9ea" containerName="console" containerID="cri-o://c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e" gracePeriod=15 Oct 08 14:14:58 crc kubenswrapper[4789]: I1008 14:14:58.866846 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm59g\" (UniqueName: \"kubernetes.io/projected/5084833a-0245-4282-8816-9d722342b260-kube-api-access-fm59g\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.054348 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.496227 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2"] Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.646166 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-dwjgm_6fd06952-1b3e-44e6-af19-dfa31965b9ea/console/0.log" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.646573 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.658267 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-service-ca\") pod \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.658358 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-oauth-config\") pod \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.658380 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwjzz\" (UniqueName: \"kubernetes.io/projected/6fd06952-1b3e-44e6-af19-dfa31965b9ea-kube-api-access-zwjzz\") pod \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.658433 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-config\") pod \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.658481 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-trusted-ca-bundle\") pod \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.658503 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-oauth-serving-cert\") pod \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.658524 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-serving-cert\") pod \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\" (UID: \"6fd06952-1b3e-44e6-af19-dfa31965b9ea\") " Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.659173 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-service-ca" (OuterVolumeSpecName: "service-ca") pod "6fd06952-1b3e-44e6-af19-dfa31965b9ea" (UID: "6fd06952-1b3e-44e6-af19-dfa31965b9ea"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.659719 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-config" (OuterVolumeSpecName: "console-config") pod "6fd06952-1b3e-44e6-af19-dfa31965b9ea" (UID: "6fd06952-1b3e-44e6-af19-dfa31965b9ea"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.660665 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6fd06952-1b3e-44e6-af19-dfa31965b9ea" (UID: "6fd06952-1b3e-44e6-af19-dfa31965b9ea"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.660824 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6fd06952-1b3e-44e6-af19-dfa31965b9ea" (UID: "6fd06952-1b3e-44e6-af19-dfa31965b9ea"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.665634 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fd06952-1b3e-44e6-af19-dfa31965b9ea-kube-api-access-zwjzz" (OuterVolumeSpecName: "kube-api-access-zwjzz") pod "6fd06952-1b3e-44e6-af19-dfa31965b9ea" (UID: "6fd06952-1b3e-44e6-af19-dfa31965b9ea"). InnerVolumeSpecName "kube-api-access-zwjzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.665687 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6fd06952-1b3e-44e6-af19-dfa31965b9ea" (UID: "6fd06952-1b3e-44e6-af19-dfa31965b9ea"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.668278 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6fd06952-1b3e-44e6-af19-dfa31965b9ea" (UID: "6fd06952-1b3e-44e6-af19-dfa31965b9ea"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.759855 4789 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.759901 4789 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.759958 4789 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.760027 4789 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.760088 4789 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.760101 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwjzz\" (UniqueName: \"kubernetes.io/projected/6fd06952-1b3e-44e6-af19-dfa31965b9ea-kube-api-access-zwjzz\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.760113 4789 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6fd06952-1b3e-44e6-af19-dfa31965b9ea-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.961863 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-dwjgm_6fd06952-1b3e-44e6-af19-dfa31965b9ea/console/0.log" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.961916 4789 generic.go:334] "Generic (PLEG): container finished" podID="6fd06952-1b3e-44e6-af19-dfa31965b9ea" containerID="c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e" exitCode=2 Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.962002 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dwjgm" event={"ID":"6fd06952-1b3e-44e6-af19-dfa31965b9ea","Type":"ContainerDied","Data":"c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e"} Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.962026 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dwjgm" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.962095 4789 scope.go:117] "RemoveContainer" containerID="c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.962246 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dwjgm" event={"ID":"6fd06952-1b3e-44e6-af19-dfa31965b9ea","Type":"ContainerDied","Data":"7b1a261d86306329063009babd8ea59e76f366a02454369d3eb7b04f06126653"} Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.963701 4789 generic.go:334] "Generic (PLEG): container finished" podID="5084833a-0245-4282-8816-9d722342b260" containerID="b3adcf06f63d43b5b506be06676a1f186ca3524921b476180f7bf65068a6c148" exitCode=0 Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.963783 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" event={"ID":"5084833a-0245-4282-8816-9d722342b260","Type":"ContainerDied","Data":"b3adcf06f63d43b5b506be06676a1f186ca3524921b476180f7bf65068a6c148"} Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.964125 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2dgnw" podUID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerName="registry-server" containerID="cri-o://2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052" gracePeriod=2 Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.965606 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" event={"ID":"5084833a-0245-4282-8816-9d722342b260","Type":"ContainerStarted","Data":"880d61896c48c2e2019579084e72fd520399aee7a8032d33552bae5f872391dc"} Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.993099 4789 scope.go:117] "RemoveContainer" containerID="c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e" Oct 08 14:14:59 crc kubenswrapper[4789]: E1008 14:14:59.995489 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e\": container with ID starting with c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e not found: ID does not exist" containerID="c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e" Oct 08 14:14:59 crc kubenswrapper[4789]: I1008 14:14:59.995534 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e"} err="failed to get container status \"c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e\": rpc error: code = NotFound desc = could not find container \"c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e\": container with ID starting with c8e3cd848e54c5bad1d44fefaccf036b57bd8158861dda55f51a42600dffd33e not found: ID does not exist" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.013073 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-dwjgm"] Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.018633 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-dwjgm"] Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.134758 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv"] Oct 08 14:15:00 crc kubenswrapper[4789]: E1008 14:15:00.134982 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd06952-1b3e-44e6-af19-dfa31965b9ea" containerName="console" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.135011 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd06952-1b3e-44e6-af19-dfa31965b9ea" containerName="console" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.135119 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fd06952-1b3e-44e6-af19-dfa31965b9ea" containerName="console" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.135541 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.145711 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.145744 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.163366 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv"] Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.165173 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpqxj\" (UniqueName: \"kubernetes.io/projected/ee6a88e2-3621-4c9f-8902-e13dd3799854-kube-api-access-vpqxj\") pod \"collect-profiles-29332215-qjnsv\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.165254 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6a88e2-3621-4c9f-8902-e13dd3799854-config-volume\") pod \"collect-profiles-29332215-qjnsv\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.165283 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6a88e2-3621-4c9f-8902-e13dd3799854-secret-volume\") pod \"collect-profiles-29332215-qjnsv\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.266130 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6a88e2-3621-4c9f-8902-e13dd3799854-config-volume\") pod \"collect-profiles-29332215-qjnsv\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.266190 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6a88e2-3621-4c9f-8902-e13dd3799854-secret-volume\") pod \"collect-profiles-29332215-qjnsv\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.266253 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpqxj\" (UniqueName: \"kubernetes.io/projected/ee6a88e2-3621-4c9f-8902-e13dd3799854-kube-api-access-vpqxj\") pod \"collect-profiles-29332215-qjnsv\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.267137 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6a88e2-3621-4c9f-8902-e13dd3799854-config-volume\") pod \"collect-profiles-29332215-qjnsv\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.280386 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6a88e2-3621-4c9f-8902-e13dd3799854-secret-volume\") pod \"collect-profiles-29332215-qjnsv\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.283981 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpqxj\" (UniqueName: \"kubernetes.io/projected/ee6a88e2-3621-4c9f-8902-e13dd3799854-kube-api-access-vpqxj\") pod \"collect-profiles-29332215-qjnsv\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.349168 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.456741 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.469032 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-utilities\") pod \"82dd900f-00ff-4626-b185-6101ea04a6fd\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.469185 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldr98\" (UniqueName: \"kubernetes.io/projected/82dd900f-00ff-4626-b185-6101ea04a6fd-kube-api-access-ldr98\") pod \"82dd900f-00ff-4626-b185-6101ea04a6fd\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.469317 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-catalog-content\") pod \"82dd900f-00ff-4626-b185-6101ea04a6fd\" (UID: \"82dd900f-00ff-4626-b185-6101ea04a6fd\") " Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.470679 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-utilities" (OuterVolumeSpecName: "utilities") pod "82dd900f-00ff-4626-b185-6101ea04a6fd" (UID: "82dd900f-00ff-4626-b185-6101ea04a6fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.473731 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82dd900f-00ff-4626-b185-6101ea04a6fd-kube-api-access-ldr98" (OuterVolumeSpecName: "kube-api-access-ldr98") pod "82dd900f-00ff-4626-b185-6101ea04a6fd" (UID: "82dd900f-00ff-4626-b185-6101ea04a6fd"). InnerVolumeSpecName "kube-api-access-ldr98". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.574413 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.574499 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldr98\" (UniqueName: \"kubernetes.io/projected/82dd900f-00ff-4626-b185-6101ea04a6fd-kube-api-access-ldr98\") on node \"crc\" DevicePath \"\"" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.745361 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fd06952-1b3e-44e6-af19-dfa31965b9ea" path="/var/lib/kubelet/pods/6fd06952-1b3e-44e6-af19-dfa31965b9ea/volumes" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.855626 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv"] Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.981492 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" event={"ID":"ee6a88e2-3621-4c9f-8902-e13dd3799854","Type":"ContainerStarted","Data":"fdb696e40e8884bfefc6e191f2179c09c578aa2be30c815442410b0557e881b5"} Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.998438 4789 generic.go:334] "Generic (PLEG): container finished" podID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerID="2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052" exitCode=0 Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.998539 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2dgnw" event={"ID":"82dd900f-00ff-4626-b185-6101ea04a6fd","Type":"ContainerDied","Data":"2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052"} Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.998573 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2dgnw" event={"ID":"82dd900f-00ff-4626-b185-6101ea04a6fd","Type":"ContainerDied","Data":"23121469d6c525b8bc55fea67d9ed2545e3ff7eaccf1150480fbe08345841394"} Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.998593 4789 scope.go:117] "RemoveContainer" containerID="2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052" Oct 08 14:15:00 crc kubenswrapper[4789]: I1008 14:15:00.998947 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2dgnw" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.031846 4789 scope.go:117] "RemoveContainer" containerID="b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.070768 4789 scope.go:117] "RemoveContainer" containerID="051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.090250 4789 scope.go:117] "RemoveContainer" containerID="2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052" Oct 08 14:15:01 crc kubenswrapper[4789]: E1008 14:15:01.090907 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052\": container with ID starting with 2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052 not found: ID does not exist" containerID="2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.090929 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052"} err="failed to get container status \"2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052\": rpc error: code = NotFound desc = could not find container \"2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052\": container with ID starting with 2739ac1a383ea57f25336e5670ed92375b7313febb61ce1e46e4bd09edb50052 not found: ID does not exist" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.090947 4789 scope.go:117] "RemoveContainer" containerID="b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac" Oct 08 14:15:01 crc kubenswrapper[4789]: E1008 14:15:01.091634 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac\": container with ID starting with b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac not found: ID does not exist" containerID="b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.091653 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac"} err="failed to get container status \"b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac\": rpc error: code = NotFound desc = could not find container \"b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac\": container with ID starting with b80756baffd63afa55ec66da442767c95c6858e279efbf20ba5c9358ec9d39ac not found: ID does not exist" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.091665 4789 scope.go:117] "RemoveContainer" containerID="051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4" Oct 08 14:15:01 crc kubenswrapper[4789]: E1008 14:15:01.091828 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4\": container with ID starting with 051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4 not found: ID does not exist" containerID="051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.091843 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4"} err="failed to get container status \"051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4\": rpc error: code = NotFound desc = could not find container \"051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4\": container with ID starting with 051314ab7018cfa436b75a2d21cdbf4eb17005be530b6427e53f78a56b8e25c4 not found: ID does not exist" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.904032 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82dd900f-00ff-4626-b185-6101ea04a6fd" (UID: "82dd900f-00ff-4626-b185-6101ea04a6fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:15:01 crc kubenswrapper[4789]: I1008 14:15:01.997621 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82dd900f-00ff-4626-b185-6101ea04a6fd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:15:02 crc kubenswrapper[4789]: I1008 14:15:02.013599 4789 generic.go:334] "Generic (PLEG): container finished" podID="ee6a88e2-3621-4c9f-8902-e13dd3799854" containerID="6a1b81d36056cbf14c53b74c5966aa4be8e99e33ade58c21a37ce6051b78cb71" exitCode=0 Oct 08 14:15:02 crc kubenswrapper[4789]: I1008 14:15:02.013644 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" event={"ID":"ee6a88e2-3621-4c9f-8902-e13dd3799854","Type":"ContainerDied","Data":"6a1b81d36056cbf14c53b74c5966aa4be8e99e33ade58c21a37ce6051b78cb71"} Oct 08 14:15:02 crc kubenswrapper[4789]: I1008 14:15:02.221893 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2dgnw"] Oct 08 14:15:02 crc kubenswrapper[4789]: I1008 14:15:02.228294 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2dgnw"] Oct 08 14:15:02 crc kubenswrapper[4789]: I1008 14:15:02.737608 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82dd900f-00ff-4626-b185-6101ea04a6fd" path="/var/lib/kubelet/pods/82dd900f-00ff-4626-b185-6101ea04a6fd/volumes" Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.020319 4789 generic.go:334] "Generic (PLEG): container finished" podID="5084833a-0245-4282-8816-9d722342b260" containerID="7a41f5cc48dc2fc42a9d0ab311ff98b11e5815160885bc55ffc3359fb6a1e34f" exitCode=0 Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.020384 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" event={"ID":"5084833a-0245-4282-8816-9d722342b260","Type":"ContainerDied","Data":"7a41f5cc48dc2fc42a9d0ab311ff98b11e5815160885bc55ffc3359fb6a1e34f"} Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.286359 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.313490 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6a88e2-3621-4c9f-8902-e13dd3799854-config-volume\") pod \"ee6a88e2-3621-4c9f-8902-e13dd3799854\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.313561 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6a88e2-3621-4c9f-8902-e13dd3799854-secret-volume\") pod \"ee6a88e2-3621-4c9f-8902-e13dd3799854\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.313600 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpqxj\" (UniqueName: \"kubernetes.io/projected/ee6a88e2-3621-4c9f-8902-e13dd3799854-kube-api-access-vpqxj\") pod \"ee6a88e2-3621-4c9f-8902-e13dd3799854\" (UID: \"ee6a88e2-3621-4c9f-8902-e13dd3799854\") " Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.318871 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee6a88e2-3621-4c9f-8902-e13dd3799854-kube-api-access-vpqxj" (OuterVolumeSpecName: "kube-api-access-vpqxj") pod "ee6a88e2-3621-4c9f-8902-e13dd3799854" (UID: "ee6a88e2-3621-4c9f-8902-e13dd3799854"). InnerVolumeSpecName "kube-api-access-vpqxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.320322 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee6a88e2-3621-4c9f-8902-e13dd3799854-config-volume" (OuterVolumeSpecName: "config-volume") pod "ee6a88e2-3621-4c9f-8902-e13dd3799854" (UID: "ee6a88e2-3621-4c9f-8902-e13dd3799854"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.323851 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee6a88e2-3621-4c9f-8902-e13dd3799854-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ee6a88e2-3621-4c9f-8902-e13dd3799854" (UID: "ee6a88e2-3621-4c9f-8902-e13dd3799854"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.414710 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee6a88e2-3621-4c9f-8902-e13dd3799854-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.414758 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee6a88e2-3621-4c9f-8902-e13dd3799854-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 14:15:03 crc kubenswrapper[4789]: I1008 14:15:03.414769 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpqxj\" (UniqueName: \"kubernetes.io/projected/ee6a88e2-3621-4c9f-8902-e13dd3799854-kube-api-access-vpqxj\") on node \"crc\" DevicePath \"\"" Oct 08 14:15:04 crc kubenswrapper[4789]: I1008 14:15:04.028428 4789 generic.go:334] "Generic (PLEG): container finished" podID="5084833a-0245-4282-8816-9d722342b260" containerID="b60b33dd976ac673bba77c16978d4ebc2c2dd342e696920cca9a86981807b16a" exitCode=0 Oct 08 14:15:04 crc kubenswrapper[4789]: I1008 14:15:04.028471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" event={"ID":"5084833a-0245-4282-8816-9d722342b260","Type":"ContainerDied","Data":"b60b33dd976ac673bba77c16978d4ebc2c2dd342e696920cca9a86981807b16a"} Oct 08 14:15:04 crc kubenswrapper[4789]: I1008 14:15:04.030094 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" event={"ID":"ee6a88e2-3621-4c9f-8902-e13dd3799854","Type":"ContainerDied","Data":"fdb696e40e8884bfefc6e191f2179c09c578aa2be30c815442410b0557e881b5"} Oct 08 14:15:04 crc kubenswrapper[4789]: I1008 14:15:04.030117 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv" Oct 08 14:15:04 crc kubenswrapper[4789]: I1008 14:15:04.030122 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdb696e40e8884bfefc6e191f2179c09c578aa2be30c815442410b0557e881b5" Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.336176 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.479054 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-util\") pod \"5084833a-0245-4282-8816-9d722342b260\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.479188 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm59g\" (UniqueName: \"kubernetes.io/projected/5084833a-0245-4282-8816-9d722342b260-kube-api-access-fm59g\") pod \"5084833a-0245-4282-8816-9d722342b260\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.479225 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-bundle\") pod \"5084833a-0245-4282-8816-9d722342b260\" (UID: \"5084833a-0245-4282-8816-9d722342b260\") " Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.480113 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-bundle" (OuterVolumeSpecName: "bundle") pod "5084833a-0245-4282-8816-9d722342b260" (UID: "5084833a-0245-4282-8816-9d722342b260"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.485589 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5084833a-0245-4282-8816-9d722342b260-kube-api-access-fm59g" (OuterVolumeSpecName: "kube-api-access-fm59g") pod "5084833a-0245-4282-8816-9d722342b260" (UID: "5084833a-0245-4282-8816-9d722342b260"). InnerVolumeSpecName "kube-api-access-fm59g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.580779 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm59g\" (UniqueName: \"kubernetes.io/projected/5084833a-0245-4282-8816-9d722342b260-kube-api-access-fm59g\") on node \"crc\" DevicePath \"\"" Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.580820 4789 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.758228 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-util" (OuterVolumeSpecName: "util") pod "5084833a-0245-4282-8816-9d722342b260" (UID: "5084833a-0245-4282-8816-9d722342b260"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:15:05 crc kubenswrapper[4789]: I1008 14:15:05.783324 4789 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5084833a-0245-4282-8816-9d722342b260-util\") on node \"crc\" DevicePath \"\"" Oct 08 14:15:06 crc kubenswrapper[4789]: I1008 14:15:06.056899 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" event={"ID":"5084833a-0245-4282-8816-9d722342b260","Type":"ContainerDied","Data":"880d61896c48c2e2019579084e72fd520399aee7a8032d33552bae5f872391dc"} Oct 08 14:15:06 crc kubenswrapper[4789]: I1008 14:15:06.056951 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="880d61896c48c2e2019579084e72fd520399aee7a8032d33552bae5f872391dc" Oct 08 14:15:06 crc kubenswrapper[4789]: I1008 14:15:06.056980 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.566035 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v"] Oct 08 14:15:17 crc kubenswrapper[4789]: E1008 14:15:17.566805 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5084833a-0245-4282-8816-9d722342b260" containerName="pull" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.566819 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5084833a-0245-4282-8816-9d722342b260" containerName="pull" Oct 08 14:15:17 crc kubenswrapper[4789]: E1008 14:15:17.566832 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerName="extract-utilities" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.566839 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerName="extract-utilities" Oct 08 14:15:17 crc kubenswrapper[4789]: E1008 14:15:17.566854 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerName="registry-server" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.566861 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerName="registry-server" Oct 08 14:15:17 crc kubenswrapper[4789]: E1008 14:15:17.566870 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5084833a-0245-4282-8816-9d722342b260" containerName="util" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.566877 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5084833a-0245-4282-8816-9d722342b260" containerName="util" Oct 08 14:15:17 crc kubenswrapper[4789]: E1008 14:15:17.566888 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerName="extract-content" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.566894 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerName="extract-content" Oct 08 14:15:17 crc kubenswrapper[4789]: E1008 14:15:17.566906 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee6a88e2-3621-4c9f-8902-e13dd3799854" containerName="collect-profiles" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.566912 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee6a88e2-3621-4c9f-8902-e13dd3799854" containerName="collect-profiles" Oct 08 14:15:17 crc kubenswrapper[4789]: E1008 14:15:17.566922 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5084833a-0245-4282-8816-9d722342b260" containerName="extract" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.566955 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5084833a-0245-4282-8816-9d722342b260" containerName="extract" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.567071 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="82dd900f-00ff-4626-b185-6101ea04a6fd" containerName="registry-server" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.567087 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee6a88e2-3621-4c9f-8902-e13dd3799854" containerName="collect-profiles" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.567095 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5084833a-0245-4282-8816-9d722342b260" containerName="extract" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.567464 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.569460 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.569679 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.569734 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-w5ksc" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.569937 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.570616 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.587456 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v"] Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.750723 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/222ee176-36f5-4f7e-b6cf-8c526a0008d6-webhook-cert\") pod \"metallb-operator-controller-manager-69d56bf45b-9wq7v\" (UID: \"222ee176-36f5-4f7e-b6cf-8c526a0008d6\") " pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.750869 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvbc7\" (UniqueName: \"kubernetes.io/projected/222ee176-36f5-4f7e-b6cf-8c526a0008d6-kube-api-access-bvbc7\") pod \"metallb-operator-controller-manager-69d56bf45b-9wq7v\" (UID: \"222ee176-36f5-4f7e-b6cf-8c526a0008d6\") " pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.750924 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/222ee176-36f5-4f7e-b6cf-8c526a0008d6-apiservice-cert\") pod \"metallb-operator-controller-manager-69d56bf45b-9wq7v\" (UID: \"222ee176-36f5-4f7e-b6cf-8c526a0008d6\") " pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.797617 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv"] Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.798514 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.801414 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.801594 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.801769 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8ms8q" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.814125 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv"] Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.852021 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd-webhook-cert\") pod \"metallb-operator-webhook-server-655bc78868-jzwhv\" (UID: \"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd\") " pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.852076 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd-apiservice-cert\") pod \"metallb-operator-webhook-server-655bc78868-jzwhv\" (UID: \"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd\") " pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.852111 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvbc7\" (UniqueName: \"kubernetes.io/projected/222ee176-36f5-4f7e-b6cf-8c526a0008d6-kube-api-access-bvbc7\") pod \"metallb-operator-controller-manager-69d56bf45b-9wq7v\" (UID: \"222ee176-36f5-4f7e-b6cf-8c526a0008d6\") " pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.852147 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/222ee176-36f5-4f7e-b6cf-8c526a0008d6-apiservice-cert\") pod \"metallb-operator-controller-manager-69d56bf45b-9wq7v\" (UID: \"222ee176-36f5-4f7e-b6cf-8c526a0008d6\") " pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.852187 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/222ee176-36f5-4f7e-b6cf-8c526a0008d6-webhook-cert\") pod \"metallb-operator-controller-manager-69d56bf45b-9wq7v\" (UID: \"222ee176-36f5-4f7e-b6cf-8c526a0008d6\") " pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.852219 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg9lg\" (UniqueName: \"kubernetes.io/projected/067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd-kube-api-access-bg9lg\") pod \"metallb-operator-webhook-server-655bc78868-jzwhv\" (UID: \"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd\") " pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.857536 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/222ee176-36f5-4f7e-b6cf-8c526a0008d6-webhook-cert\") pod \"metallb-operator-controller-manager-69d56bf45b-9wq7v\" (UID: \"222ee176-36f5-4f7e-b6cf-8c526a0008d6\") " pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.857607 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/222ee176-36f5-4f7e-b6cf-8c526a0008d6-apiservice-cert\") pod \"metallb-operator-controller-manager-69d56bf45b-9wq7v\" (UID: \"222ee176-36f5-4f7e-b6cf-8c526a0008d6\") " pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.868697 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvbc7\" (UniqueName: \"kubernetes.io/projected/222ee176-36f5-4f7e-b6cf-8c526a0008d6-kube-api-access-bvbc7\") pod \"metallb-operator-controller-manager-69d56bf45b-9wq7v\" (UID: \"222ee176-36f5-4f7e-b6cf-8c526a0008d6\") " pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.890637 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.952934 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg9lg\" (UniqueName: \"kubernetes.io/projected/067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd-kube-api-access-bg9lg\") pod \"metallb-operator-webhook-server-655bc78868-jzwhv\" (UID: \"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd\") " pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.953024 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd-webhook-cert\") pod \"metallb-operator-webhook-server-655bc78868-jzwhv\" (UID: \"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd\") " pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.953049 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd-apiservice-cert\") pod \"metallb-operator-webhook-server-655bc78868-jzwhv\" (UID: \"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd\") " pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.956683 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd-apiservice-cert\") pod \"metallb-operator-webhook-server-655bc78868-jzwhv\" (UID: \"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd\") " pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.956722 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd-webhook-cert\") pod \"metallb-operator-webhook-server-655bc78868-jzwhv\" (UID: \"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd\") " pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:17 crc kubenswrapper[4789]: I1008 14:15:17.973960 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg9lg\" (UniqueName: \"kubernetes.io/projected/067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd-kube-api-access-bg9lg\") pod \"metallb-operator-webhook-server-655bc78868-jzwhv\" (UID: \"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd\") " pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:18 crc kubenswrapper[4789]: I1008 14:15:18.111322 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:18 crc kubenswrapper[4789]: I1008 14:15:18.340574 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v"] Oct 08 14:15:18 crc kubenswrapper[4789]: I1008 14:15:18.560728 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv"] Oct 08 14:15:18 crc kubenswrapper[4789]: W1008 14:15:18.568114 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod067b91bf_82a6_4d6d_bf91_6f0cf31fa6dd.slice/crio-b7e6d3da0aca97ee579e3dce04f74754c625da60c8b115df8019f19f2673f45a WatchSource:0}: Error finding container b7e6d3da0aca97ee579e3dce04f74754c625da60c8b115df8019f19f2673f45a: Status 404 returned error can't find the container with id b7e6d3da0aca97ee579e3dce04f74754c625da60c8b115df8019f19f2673f45a Oct 08 14:15:19 crc kubenswrapper[4789]: I1008 14:15:19.148846 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" event={"ID":"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd","Type":"ContainerStarted","Data":"b7e6d3da0aca97ee579e3dce04f74754c625da60c8b115df8019f19f2673f45a"} Oct 08 14:15:19 crc kubenswrapper[4789]: I1008 14:15:19.150661 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" event={"ID":"222ee176-36f5-4f7e-b6cf-8c526a0008d6","Type":"ContainerStarted","Data":"207884675fa9a9f145119f8d9eecb4a302d8a6ae44074dac389978b1e0963f82"} Oct 08 14:15:26 crc kubenswrapper[4789]: I1008 14:15:26.191844 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" event={"ID":"222ee176-36f5-4f7e-b6cf-8c526a0008d6","Type":"ContainerStarted","Data":"07340dfd918b140afe86a4b71428e2a2bb9998bd557f37a89bb44cc5dcc79530"} Oct 08 14:15:26 crc kubenswrapper[4789]: I1008 14:15:26.192510 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:26 crc kubenswrapper[4789]: I1008 14:15:26.194844 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" event={"ID":"067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd","Type":"ContainerStarted","Data":"76d76476af2df817a0ec7afafa00167486efbb1b0c34703e7d942373f7d61dfd"} Oct 08 14:15:26 crc kubenswrapper[4789]: I1008 14:15:26.195287 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:26 crc kubenswrapper[4789]: I1008 14:15:26.224275 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" podStartSLOduration=1.896078988 podStartE2EDuration="9.224257873s" podCreationTimestamp="2025-10-08 14:15:17 +0000 UTC" firstStartedPulling="2025-10-08 14:15:18.342280035 +0000 UTC m=+858.249027527" lastFinishedPulling="2025-10-08 14:15:25.67045892 +0000 UTC m=+865.577206412" observedRunningTime="2025-10-08 14:15:26.220313426 +0000 UTC m=+866.127060928" watchObservedRunningTime="2025-10-08 14:15:26.224257873 +0000 UTC m=+866.131005375" Oct 08 14:15:26 crc kubenswrapper[4789]: I1008 14:15:26.246497 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" podStartSLOduration=2.131925821 podStartE2EDuration="9.246470575s" podCreationTimestamp="2025-10-08 14:15:17 +0000 UTC" firstStartedPulling="2025-10-08 14:15:18.571721685 +0000 UTC m=+858.478469177" lastFinishedPulling="2025-10-08 14:15:25.686266439 +0000 UTC m=+865.593013931" observedRunningTime="2025-10-08 14:15:26.245772856 +0000 UTC m=+866.152520358" watchObservedRunningTime="2025-10-08 14:15:26.246470575 +0000 UTC m=+866.153218107" Oct 08 14:15:38 crc kubenswrapper[4789]: I1008 14:15:38.119720 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-655bc78868-jzwhv" Oct 08 14:15:56 crc kubenswrapper[4789]: I1008 14:15:56.432775 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:15:56 crc kubenswrapper[4789]: I1008 14:15:56.433480 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:15:57 crc kubenswrapper[4789]: I1008 14:15:57.893495 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-69d56bf45b-9wq7v" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.569971 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-bq6zf"] Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.572926 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.575484 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-jtlrk" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.575684 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.576435 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.585051 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs"] Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.586246 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.588156 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.598027 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs"] Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.668811 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-646mh"] Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.669748 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.673889 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.674170 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.674542 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.674730 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dzfcf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.680827 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-ggzfl"] Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.681794 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.684444 4789 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.696106 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-ggzfl"] Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.734547 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-reloader\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.734595 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-metrics\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.734632 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-frr-conf\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.734659 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-metrics-certs\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.734689 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/988fe042-8dd6-4398-af3d-ae201442f223-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qx7cs\" (UID: \"988fe042-8dd6-4398-af3d-ae201442f223\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.734773 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8jp9\" (UniqueName: \"kubernetes.io/projected/988fe042-8dd6-4398-af3d-ae201442f223-kube-api-access-b8jp9\") pod \"frr-k8s-webhook-server-64bf5d555-qx7cs\" (UID: \"988fe042-8dd6-4398-af3d-ae201442f223\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.734925 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shl9j\" (UniqueName: \"kubernetes.io/projected/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-kube-api-access-shl9j\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.734980 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-frr-sockets\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.735025 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-frr-startup\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.836646 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-metrics-certs\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.836695 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/988fe042-8dd6-4398-af3d-ae201442f223-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qx7cs\" (UID: \"988fe042-8dd6-4398-af3d-ae201442f223\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.836729 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f942609-b5dc-4433-8767-a9bd36b185f9-metrics-certs\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.836762 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8jp9\" (UniqueName: \"kubernetes.io/projected/988fe042-8dd6-4398-af3d-ae201442f223-kube-api-access-b8jp9\") pod \"frr-k8s-webhook-server-64bf5d555-qx7cs\" (UID: \"988fe042-8dd6-4398-af3d-ae201442f223\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.836822 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkfcf\" (UniqueName: \"kubernetes.io/projected/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-kube-api-access-xkfcf\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.836849 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-metallb-excludel2\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.836875 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shl9j\" (UniqueName: \"kubernetes.io/projected/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-kube-api-access-shl9j\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: E1008 14:15:58.836880 4789 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.836902 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-frr-sockets\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: E1008 14:15:58.836965 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/988fe042-8dd6-4398-af3d-ae201442f223-cert podName:988fe042-8dd6-4398-af3d-ae201442f223 nodeName:}" failed. No retries permitted until 2025-10-08 14:15:59.336941108 +0000 UTC m=+899.243688600 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/988fe042-8dd6-4398-af3d-ae201442f223-cert") pod "frr-k8s-webhook-server-64bf5d555-qx7cs" (UID: "988fe042-8dd6-4398-af3d-ae201442f223") : secret "frr-k8s-webhook-server-cert" not found Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837039 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-metrics-certs\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837075 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-frr-startup\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837127 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f942609-b5dc-4433-8767-a9bd36b185f9-cert\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837155 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-reloader\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837184 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-metrics\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837269 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxkjj\" (UniqueName: \"kubernetes.io/projected/9f942609-b5dc-4433-8767-a9bd36b185f9-kube-api-access-gxkjj\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837288 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-frr-sockets\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837302 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-memberlist\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837387 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-frr-conf\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837495 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-metrics\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837512 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-reloader\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.837716 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-frr-conf\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.838172 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-frr-startup\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.855708 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-metrics-certs\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.866154 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8jp9\" (UniqueName: \"kubernetes.io/projected/988fe042-8dd6-4398-af3d-ae201442f223-kube-api-access-b8jp9\") pod \"frr-k8s-webhook-server-64bf5d555-qx7cs\" (UID: \"988fe042-8dd6-4398-af3d-ae201442f223\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.866197 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shl9j\" (UniqueName: \"kubernetes.io/projected/56b10d37-ef4c-4e34-87d1-2c15d04d63bc-kube-api-access-shl9j\") pod \"frr-k8s-bq6zf\" (UID: \"56b10d37-ef4c-4e34-87d1-2c15d04d63bc\") " pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.892765 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.939454 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkfcf\" (UniqueName: \"kubernetes.io/projected/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-kube-api-access-xkfcf\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.939515 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-metallb-excludel2\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.939544 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-metrics-certs\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.939570 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f942609-b5dc-4433-8767-a9bd36b185f9-cert\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.939608 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxkjj\" (UniqueName: \"kubernetes.io/projected/9f942609-b5dc-4433-8767-a9bd36b185f9-kube-api-access-gxkjj\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.939623 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-memberlist\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.939669 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f942609-b5dc-4433-8767-a9bd36b185f9-metrics-certs\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:58 crc kubenswrapper[4789]: E1008 14:15:58.939788 4789 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 08 14:15:58 crc kubenswrapper[4789]: E1008 14:15:58.939838 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f942609-b5dc-4433-8767-a9bd36b185f9-metrics-certs podName:9f942609-b5dc-4433-8767-a9bd36b185f9 nodeName:}" failed. No retries permitted until 2025-10-08 14:15:59.439823386 +0000 UTC m=+899.346570878 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9f942609-b5dc-4433-8767-a9bd36b185f9-metrics-certs") pod "controller-68d546b9d8-ggzfl" (UID: "9f942609-b5dc-4433-8767-a9bd36b185f9") : secret "controller-certs-secret" not found Oct 08 14:15:58 crc kubenswrapper[4789]: E1008 14:15:58.940100 4789 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 14:15:58 crc kubenswrapper[4789]: E1008 14:15:58.940128 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-memberlist podName:b76ec5f4-3801-4d1a-9b64-5b7beebc4100 nodeName:}" failed. No retries permitted until 2025-10-08 14:15:59.440119424 +0000 UTC m=+899.346866916 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-memberlist") pod "speaker-646mh" (UID: "b76ec5f4-3801-4d1a-9b64-5b7beebc4100") : secret "metallb-memberlist" not found Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.940347 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-metallb-excludel2\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.943209 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-metrics-certs\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.945754 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f942609-b5dc-4433-8767-a9bd36b185f9-cert\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.956437 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxkjj\" (UniqueName: \"kubernetes.io/projected/9f942609-b5dc-4433-8767-a9bd36b185f9-kube-api-access-gxkjj\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:58 crc kubenswrapper[4789]: I1008 14:15:58.957852 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkfcf\" (UniqueName: \"kubernetes.io/projected/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-kube-api-access-xkfcf\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.344784 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/988fe042-8dd6-4398-af3d-ae201442f223-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qx7cs\" (UID: \"988fe042-8dd6-4398-af3d-ae201442f223\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.349319 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/988fe042-8dd6-4398-af3d-ae201442f223-cert\") pod \"frr-k8s-webhook-server-64bf5d555-qx7cs\" (UID: \"988fe042-8dd6-4398-af3d-ae201442f223\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.375709 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerStarted","Data":"be5d37bd2f5b3a2bcd052c7b127f4660ad0373dd1d9dae10914362376b6b3454"} Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.446690 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-memberlist\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.446864 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f942609-b5dc-4433-8767-a9bd36b185f9-metrics-certs\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:59 crc kubenswrapper[4789]: E1008 14:15:59.446978 4789 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 14:15:59 crc kubenswrapper[4789]: E1008 14:15:59.447151 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-memberlist podName:b76ec5f4-3801-4d1a-9b64-5b7beebc4100 nodeName:}" failed. No retries permitted until 2025-10-08 14:16:00.447122879 +0000 UTC m=+900.353870371 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-memberlist") pod "speaker-646mh" (UID: "b76ec5f4-3801-4d1a-9b64-5b7beebc4100") : secret "metallb-memberlist" not found Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.450848 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f942609-b5dc-4433-8767-a9bd36b185f9-metrics-certs\") pod \"controller-68d546b9d8-ggzfl\" (UID: \"9f942609-b5dc-4433-8767-a9bd36b185f9\") " pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.510462 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.603034 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.697756 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs"] Oct 08 14:15:59 crc kubenswrapper[4789]: W1008 14:15:59.702872 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod988fe042_8dd6_4398_af3d_ae201442f223.slice/crio-30cbfed81d09d0451b60c5fe97ce0d0335996a75687b9e34e749177d5e03ae61 WatchSource:0}: Error finding container 30cbfed81d09d0451b60c5fe97ce0d0335996a75687b9e34e749177d5e03ae61: Status 404 returned error can't find the container with id 30cbfed81d09d0451b60c5fe97ce0d0335996a75687b9e34e749177d5e03ae61 Oct 08 14:15:59 crc kubenswrapper[4789]: I1008 14:15:59.817636 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-ggzfl"] Oct 08 14:15:59 crc kubenswrapper[4789]: W1008 14:15:59.822340 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f942609_b5dc_4433_8767_a9bd36b185f9.slice/crio-d84a947b1dd8f0ef76f072767221f306c50a569ad2583a8165214cbd2a53d466 WatchSource:0}: Error finding container d84a947b1dd8f0ef76f072767221f306c50a569ad2583a8165214cbd2a53d466: Status 404 returned error can't find the container with id d84a947b1dd8f0ef76f072767221f306c50a569ad2583a8165214cbd2a53d466 Oct 08 14:16:00 crc kubenswrapper[4789]: I1008 14:16:00.385255 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-ggzfl" event={"ID":"9f942609-b5dc-4433-8767-a9bd36b185f9","Type":"ContainerStarted","Data":"93603f582082e51b7e076f77f88ac72f90e6514f931910fcf0452f39f8f4f89b"} Oct 08 14:16:00 crc kubenswrapper[4789]: I1008 14:16:00.385313 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-ggzfl" event={"ID":"9f942609-b5dc-4433-8767-a9bd36b185f9","Type":"ContainerStarted","Data":"3a74112583bc541ec7b227230337490dea899c9040404b21abfd1f2264374157"} Oct 08 14:16:00 crc kubenswrapper[4789]: I1008 14:16:00.385336 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-ggzfl" event={"ID":"9f942609-b5dc-4433-8767-a9bd36b185f9","Type":"ContainerStarted","Data":"d84a947b1dd8f0ef76f072767221f306c50a569ad2583a8165214cbd2a53d466"} Oct 08 14:16:00 crc kubenswrapper[4789]: I1008 14:16:00.386061 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" event={"ID":"988fe042-8dd6-4398-af3d-ae201442f223","Type":"ContainerStarted","Data":"30cbfed81d09d0451b60c5fe97ce0d0335996a75687b9e34e749177d5e03ae61"} Oct 08 14:16:00 crc kubenswrapper[4789]: I1008 14:16:00.402740 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-ggzfl" podStartSLOduration=2.402718044 podStartE2EDuration="2.402718044s" podCreationTimestamp="2025-10-08 14:15:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:16:00.398584471 +0000 UTC m=+900.305331963" watchObservedRunningTime="2025-10-08 14:16:00.402718044 +0000 UTC m=+900.309465536" Oct 08 14:16:00 crc kubenswrapper[4789]: I1008 14:16:00.459799 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-memberlist\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:16:00 crc kubenswrapper[4789]: I1008 14:16:00.465948 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b76ec5f4-3801-4d1a-9b64-5b7beebc4100-memberlist\") pod \"speaker-646mh\" (UID: \"b76ec5f4-3801-4d1a-9b64-5b7beebc4100\") " pod="metallb-system/speaker-646mh" Oct 08 14:16:00 crc kubenswrapper[4789]: I1008 14:16:00.488707 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-646mh" Oct 08 14:16:00 crc kubenswrapper[4789]: W1008 14:16:00.520246 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb76ec5f4_3801_4d1a_9b64_5b7beebc4100.slice/crio-f885907f4bc1bffb5559e5fe6a66bb28669cc4a28cf6f013f107f074d678fe2a WatchSource:0}: Error finding container f885907f4bc1bffb5559e5fe6a66bb28669cc4a28cf6f013f107f074d678fe2a: Status 404 returned error can't find the container with id f885907f4bc1bffb5559e5fe6a66bb28669cc4a28cf6f013f107f074d678fe2a Oct 08 14:16:01 crc kubenswrapper[4789]: I1008 14:16:01.397102 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-646mh" event={"ID":"b76ec5f4-3801-4d1a-9b64-5b7beebc4100","Type":"ContainerStarted","Data":"bcfb20f98541aa5a0811be4bf56715921a0235f65e02d8ab3774f3cb87ae6c6e"} Oct 08 14:16:01 crc kubenswrapper[4789]: I1008 14:16:01.397152 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:16:01 crc kubenswrapper[4789]: I1008 14:16:01.397165 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-646mh" event={"ID":"b76ec5f4-3801-4d1a-9b64-5b7beebc4100","Type":"ContainerStarted","Data":"02d60bcde145dbfec79c381a6d727aeffb7ad741378cf26fb08bdddf018b52ee"} Oct 08 14:16:01 crc kubenswrapper[4789]: I1008 14:16:01.397178 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-646mh" event={"ID":"b76ec5f4-3801-4d1a-9b64-5b7beebc4100","Type":"ContainerStarted","Data":"f885907f4bc1bffb5559e5fe6a66bb28669cc4a28cf6f013f107f074d678fe2a"} Oct 08 14:16:01 crc kubenswrapper[4789]: I1008 14:16:01.397840 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-646mh" Oct 08 14:16:01 crc kubenswrapper[4789]: I1008 14:16:01.423173 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-646mh" podStartSLOduration=3.423152536 podStartE2EDuration="3.423152536s" podCreationTimestamp="2025-10-08 14:15:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:16:01.421292366 +0000 UTC m=+901.328039858" watchObservedRunningTime="2025-10-08 14:16:01.423152536 +0000 UTC m=+901.329900028" Oct 08 14:16:07 crc kubenswrapper[4789]: I1008 14:16:07.445025 4789 generic.go:334] "Generic (PLEG): container finished" podID="56b10d37-ef4c-4e34-87d1-2c15d04d63bc" containerID="d25fc87d9483ee624d8163c759c412b1844100a0943f9ee93d8cb6b1d1d30d39" exitCode=0 Oct 08 14:16:07 crc kubenswrapper[4789]: I1008 14:16:07.445088 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerDied","Data":"d25fc87d9483ee624d8163c759c412b1844100a0943f9ee93d8cb6b1d1d30d39"} Oct 08 14:16:07 crc kubenswrapper[4789]: I1008 14:16:07.446452 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" event={"ID":"988fe042-8dd6-4398-af3d-ae201442f223","Type":"ContainerStarted","Data":"1cf47938906ca24287fbf4df4ab12efd3ef2acc2ff862c8dd84241248123e37e"} Oct 08 14:16:07 crc kubenswrapper[4789]: I1008 14:16:07.446789 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:16:07 crc kubenswrapper[4789]: I1008 14:16:07.496235 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" podStartSLOduration=2.833754568 podStartE2EDuration="9.496217736s" podCreationTimestamp="2025-10-08 14:15:58 +0000 UTC" firstStartedPulling="2025-10-08 14:15:59.705838032 +0000 UTC m=+899.612585524" lastFinishedPulling="2025-10-08 14:16:06.3683012 +0000 UTC m=+906.275048692" observedRunningTime="2025-10-08 14:16:07.495430455 +0000 UTC m=+907.402177957" watchObservedRunningTime="2025-10-08 14:16:07.496217736 +0000 UTC m=+907.402965228" Oct 08 14:16:08 crc kubenswrapper[4789]: I1008 14:16:08.453458 4789 generic.go:334] "Generic (PLEG): container finished" podID="56b10d37-ef4c-4e34-87d1-2c15d04d63bc" containerID="6e361ef92a0385121d558293f1d2814e4ccbffcbc7ac4d047a72f1ffb2fcbfd5" exitCode=0 Oct 08 14:16:08 crc kubenswrapper[4789]: I1008 14:16:08.453551 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerDied","Data":"6e361ef92a0385121d558293f1d2814e4ccbffcbc7ac4d047a72f1ffb2fcbfd5"} Oct 08 14:16:09 crc kubenswrapper[4789]: I1008 14:16:09.460296 4789 generic.go:334] "Generic (PLEG): container finished" podID="56b10d37-ef4c-4e34-87d1-2c15d04d63bc" containerID="d97e7e35c62ad7858abb764c1dcb142418cf7fcef9d1dda1068415c82c8350d3" exitCode=0 Oct 08 14:16:09 crc kubenswrapper[4789]: I1008 14:16:09.460386 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerDied","Data":"d97e7e35c62ad7858abb764c1dcb142418cf7fcef9d1dda1068415c82c8350d3"} Oct 08 14:16:09 crc kubenswrapper[4789]: I1008 14:16:09.607044 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-ggzfl" Oct 08 14:16:10 crc kubenswrapper[4789]: I1008 14:16:10.475576 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerStarted","Data":"a76ec409dbfdbc5a662f47dd90882e17d591b368767a6d635d150d3fcfbec17f"} Oct 08 14:16:10 crc kubenswrapper[4789]: I1008 14:16:10.475864 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerStarted","Data":"7821c587010f5ed8b50a33996170c88eda29bfce7aac37c44c8d0ffdf7bdb79a"} Oct 08 14:16:10 crc kubenswrapper[4789]: I1008 14:16:10.475874 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerStarted","Data":"c877258d57e40a7ca8ce1251d06cb63b67de0455ed3e2da05d0b4abc8cfbae92"} Oct 08 14:16:10 crc kubenswrapper[4789]: I1008 14:16:10.475885 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerStarted","Data":"646d92c01ef2c42abf3a89c458361249ad37b711476b6e2646a3fceb4450432d"} Oct 08 14:16:10 crc kubenswrapper[4789]: I1008 14:16:10.493230 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-646mh" Oct 08 14:16:11 crc kubenswrapper[4789]: I1008 14:16:11.486757 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerStarted","Data":"d51a9255e42692c9c7674f131f812ff9af6bedbc62d0004809bfcb0515f5f67d"} Oct 08 14:16:11 crc kubenswrapper[4789]: I1008 14:16:11.487111 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bq6zf" event={"ID":"56b10d37-ef4c-4e34-87d1-2c15d04d63bc","Type":"ContainerStarted","Data":"57fac6308642a697e5607761a4f9001e143f3219c9630ccccd20bfe099b08ccd"} Oct 08 14:16:11 crc kubenswrapper[4789]: I1008 14:16:11.487135 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:16:11 crc kubenswrapper[4789]: I1008 14:16:11.508539 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-bq6zf" podStartSLOduration=6.163248102 podStartE2EDuration="13.50852272s" podCreationTimestamp="2025-10-08 14:15:58 +0000 UTC" firstStartedPulling="2025-10-08 14:15:59.006633498 +0000 UTC m=+898.913380990" lastFinishedPulling="2025-10-08 14:16:06.351908116 +0000 UTC m=+906.258655608" observedRunningTime="2025-10-08 14:16:11.505249251 +0000 UTC m=+911.411996743" watchObservedRunningTime="2025-10-08 14:16:11.50852272 +0000 UTC m=+911.415270212" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.545696 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-kcdv7"] Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.546887 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kcdv7" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.549094 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-x228p" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.549222 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.549522 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.554875 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kcdv7"] Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.636179 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sdfp\" (UniqueName: \"kubernetes.io/projected/29a2cf10-ed5a-46d7-a891-0e15437a1a06-kube-api-access-2sdfp\") pod \"openstack-operator-index-kcdv7\" (UID: \"29a2cf10-ed5a-46d7-a891-0e15437a1a06\") " pod="openstack-operators/openstack-operator-index-kcdv7" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.737035 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sdfp\" (UniqueName: \"kubernetes.io/projected/29a2cf10-ed5a-46d7-a891-0e15437a1a06-kube-api-access-2sdfp\") pod \"openstack-operator-index-kcdv7\" (UID: \"29a2cf10-ed5a-46d7-a891-0e15437a1a06\") " pod="openstack-operators/openstack-operator-index-kcdv7" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.754695 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sdfp\" (UniqueName: \"kubernetes.io/projected/29a2cf10-ed5a-46d7-a891-0e15437a1a06-kube-api-access-2sdfp\") pod \"openstack-operator-index-kcdv7\" (UID: \"29a2cf10-ed5a-46d7-a891-0e15437a1a06\") " pod="openstack-operators/openstack-operator-index-kcdv7" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.871652 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kcdv7" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.893400 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:16:13 crc kubenswrapper[4789]: I1008 14:16:13.930411 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:16:14 crc kubenswrapper[4789]: I1008 14:16:14.289411 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kcdv7"] Oct 08 14:16:14 crc kubenswrapper[4789]: W1008 14:16:14.290913 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29a2cf10_ed5a_46d7_a891_0e15437a1a06.slice/crio-9adf4a8525e9dde45b9d6fa2bfe1e5002a5a3a919fb92097523396663a31d32e WatchSource:0}: Error finding container 9adf4a8525e9dde45b9d6fa2bfe1e5002a5a3a919fb92097523396663a31d32e: Status 404 returned error can't find the container with id 9adf4a8525e9dde45b9d6fa2bfe1e5002a5a3a919fb92097523396663a31d32e Oct 08 14:16:14 crc kubenswrapper[4789]: I1008 14:16:14.505976 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kcdv7" event={"ID":"29a2cf10-ed5a-46d7-a891-0e15437a1a06","Type":"ContainerStarted","Data":"9adf4a8525e9dde45b9d6fa2bfe1e5002a5a3a919fb92097523396663a31d32e"} Oct 08 14:16:16 crc kubenswrapper[4789]: I1008 14:16:16.934531 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kcdv7"] Oct 08 14:16:17 crc kubenswrapper[4789]: I1008 14:16:17.542777 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-swzf7"] Oct 08 14:16:17 crc kubenswrapper[4789]: I1008 14:16:17.544227 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-swzf7" Oct 08 14:16:17 crc kubenswrapper[4789]: I1008 14:16:17.549396 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-swzf7"] Oct 08 14:16:17 crc kubenswrapper[4789]: I1008 14:16:17.699556 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz8tv\" (UniqueName: \"kubernetes.io/projected/cb5f8cb4-1588-4ae9-88b1-890831151eb4-kube-api-access-rz8tv\") pod \"openstack-operator-index-swzf7\" (UID: \"cb5f8cb4-1588-4ae9-88b1-890831151eb4\") " pod="openstack-operators/openstack-operator-index-swzf7" Oct 08 14:16:17 crc kubenswrapper[4789]: I1008 14:16:17.801233 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz8tv\" (UniqueName: \"kubernetes.io/projected/cb5f8cb4-1588-4ae9-88b1-890831151eb4-kube-api-access-rz8tv\") pod \"openstack-operator-index-swzf7\" (UID: \"cb5f8cb4-1588-4ae9-88b1-890831151eb4\") " pod="openstack-operators/openstack-operator-index-swzf7" Oct 08 14:16:17 crc kubenswrapper[4789]: I1008 14:16:17.818355 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz8tv\" (UniqueName: \"kubernetes.io/projected/cb5f8cb4-1588-4ae9-88b1-890831151eb4-kube-api-access-rz8tv\") pod \"openstack-operator-index-swzf7\" (UID: \"cb5f8cb4-1588-4ae9-88b1-890831151eb4\") " pod="openstack-operators/openstack-operator-index-swzf7" Oct 08 14:16:17 crc kubenswrapper[4789]: I1008 14:16:17.865183 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-swzf7" Oct 08 14:16:19 crc kubenswrapper[4789]: I1008 14:16:19.107681 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-swzf7"] Oct 08 14:16:19 crc kubenswrapper[4789]: W1008 14:16:19.114449 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb5f8cb4_1588_4ae9_88b1_890831151eb4.slice/crio-f291abaffcb4444f7ab3d51a3bef1632cd19b7ade5eb6de2fc6caf8874b41ea5 WatchSource:0}: Error finding container f291abaffcb4444f7ab3d51a3bef1632cd19b7ade5eb6de2fc6caf8874b41ea5: Status 404 returned error can't find the container with id f291abaffcb4444f7ab3d51a3bef1632cd19b7ade5eb6de2fc6caf8874b41ea5 Oct 08 14:16:19 crc kubenswrapper[4789]: I1008 14:16:19.515782 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-qx7cs" Oct 08 14:16:19 crc kubenswrapper[4789]: I1008 14:16:19.556068 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-swzf7" event={"ID":"cb5f8cb4-1588-4ae9-88b1-890831151eb4","Type":"ContainerStarted","Data":"82724dd6d3a9064743eaebfc32dd9d4b4c7521e5adb7bc639885a27b848f9469"} Oct 08 14:16:19 crc kubenswrapper[4789]: I1008 14:16:19.556136 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-swzf7" event={"ID":"cb5f8cb4-1588-4ae9-88b1-890831151eb4","Type":"ContainerStarted","Data":"f291abaffcb4444f7ab3d51a3bef1632cd19b7ade5eb6de2fc6caf8874b41ea5"} Oct 08 14:16:19 crc kubenswrapper[4789]: I1008 14:16:19.557744 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kcdv7" event={"ID":"29a2cf10-ed5a-46d7-a891-0e15437a1a06","Type":"ContainerStarted","Data":"450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34"} Oct 08 14:16:19 crc kubenswrapper[4789]: I1008 14:16:19.557875 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-kcdv7" podUID="29a2cf10-ed5a-46d7-a891-0e15437a1a06" containerName="registry-server" containerID="cri-o://450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34" gracePeriod=2 Oct 08 14:16:19 crc kubenswrapper[4789]: I1008 14:16:19.581361 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-swzf7" podStartSLOduration=2.31761186 podStartE2EDuration="2.581343198s" podCreationTimestamp="2025-10-08 14:16:17 +0000 UTC" firstStartedPulling="2025-10-08 14:16:19.120888489 +0000 UTC m=+919.027635991" lastFinishedPulling="2025-10-08 14:16:19.384619797 +0000 UTC m=+919.291367329" observedRunningTime="2025-10-08 14:16:19.577577096 +0000 UTC m=+919.484324608" watchObservedRunningTime="2025-10-08 14:16:19.581343198 +0000 UTC m=+919.488090690" Oct 08 14:16:19 crc kubenswrapper[4789]: I1008 14:16:19.596167 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-kcdv7" podStartSLOduration=1.9026003249999999 podStartE2EDuration="6.596149009s" podCreationTimestamp="2025-10-08 14:16:13 +0000 UTC" firstStartedPulling="2025-10-08 14:16:14.293314543 +0000 UTC m=+914.200062075" lastFinishedPulling="2025-10-08 14:16:18.986863267 +0000 UTC m=+918.893610759" observedRunningTime="2025-10-08 14:16:19.594872235 +0000 UTC m=+919.501619727" watchObservedRunningTime="2025-10-08 14:16:19.596149009 +0000 UTC m=+919.502896501" Oct 08 14:16:19 crc kubenswrapper[4789]: I1008 14:16:19.937951 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kcdv7" Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.064260 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sdfp\" (UniqueName: \"kubernetes.io/projected/29a2cf10-ed5a-46d7-a891-0e15437a1a06-kube-api-access-2sdfp\") pod \"29a2cf10-ed5a-46d7-a891-0e15437a1a06\" (UID: \"29a2cf10-ed5a-46d7-a891-0e15437a1a06\") " Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.069502 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a2cf10-ed5a-46d7-a891-0e15437a1a06-kube-api-access-2sdfp" (OuterVolumeSpecName: "kube-api-access-2sdfp") pod "29a2cf10-ed5a-46d7-a891-0e15437a1a06" (UID: "29a2cf10-ed5a-46d7-a891-0e15437a1a06"). InnerVolumeSpecName "kube-api-access-2sdfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.165800 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sdfp\" (UniqueName: \"kubernetes.io/projected/29a2cf10-ed5a-46d7-a891-0e15437a1a06-kube-api-access-2sdfp\") on node \"crc\" DevicePath \"\"" Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.567673 4789 generic.go:334] "Generic (PLEG): container finished" podID="29a2cf10-ed5a-46d7-a891-0e15437a1a06" containerID="450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34" exitCode=0 Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.567740 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kcdv7" event={"ID":"29a2cf10-ed5a-46d7-a891-0e15437a1a06","Type":"ContainerDied","Data":"450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34"} Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.567817 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kcdv7" event={"ID":"29a2cf10-ed5a-46d7-a891-0e15437a1a06","Type":"ContainerDied","Data":"9adf4a8525e9dde45b9d6fa2bfe1e5002a5a3a919fb92097523396663a31d32e"} Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.567840 4789 scope.go:117] "RemoveContainer" containerID="450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34" Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.567750 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kcdv7" Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.588813 4789 scope.go:117] "RemoveContainer" containerID="450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34" Oct 08 14:16:20 crc kubenswrapper[4789]: E1008 14:16:20.589503 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34\": container with ID starting with 450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34 not found: ID does not exist" containerID="450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34" Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.589559 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34"} err="failed to get container status \"450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34\": rpc error: code = NotFound desc = could not find container \"450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34\": container with ID starting with 450893dc69454fb842735b5befc42e5e000d0e21956a7ea02b38abdc631e7b34 not found: ID does not exist" Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.597969 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kcdv7"] Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.608461 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-kcdv7"] Oct 08 14:16:20 crc kubenswrapper[4789]: I1008 14:16:20.738032 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29a2cf10-ed5a-46d7-a891-0e15437a1a06" path="/var/lib/kubelet/pods/29a2cf10-ed5a-46d7-a891-0e15437a1a06/volumes" Oct 08 14:16:26 crc kubenswrapper[4789]: I1008 14:16:26.433510 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:16:26 crc kubenswrapper[4789]: I1008 14:16:26.434338 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:16:27 crc kubenswrapper[4789]: I1008 14:16:27.865836 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-swzf7" Oct 08 14:16:27 crc kubenswrapper[4789]: I1008 14:16:27.866291 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-swzf7" Oct 08 14:16:27 crc kubenswrapper[4789]: I1008 14:16:27.898109 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-swzf7" Oct 08 14:16:28 crc kubenswrapper[4789]: I1008 14:16:28.670138 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-swzf7" Oct 08 14:16:28 crc kubenswrapper[4789]: I1008 14:16:28.896302 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-bq6zf" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.528828 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq"] Oct 08 14:16:35 crc kubenswrapper[4789]: E1008 14:16:35.529616 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a2cf10-ed5a-46d7-a891-0e15437a1a06" containerName="registry-server" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.529629 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a2cf10-ed5a-46d7-a891-0e15437a1a06" containerName="registry-server" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.529779 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a2cf10-ed5a-46d7-a891-0e15437a1a06" containerName="registry-server" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.530773 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.534970 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dgdgp" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.536560 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq"] Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.628720 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-util\") pod \"96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.628770 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-bundle\") pod \"96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.629007 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb2qs\" (UniqueName: \"kubernetes.io/projected/db7422df-8be7-4a79-9981-b5fd2348e4c4-kube-api-access-sb2qs\") pod \"96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.729855 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb2qs\" (UniqueName: \"kubernetes.io/projected/db7422df-8be7-4a79-9981-b5fd2348e4c4-kube-api-access-sb2qs\") pod \"96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.729932 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-util\") pod \"96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.729964 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-bundle\") pod \"96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.730584 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-util\") pod \"96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.730606 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-bundle\") pod \"96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.750276 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb2qs\" (UniqueName: \"kubernetes.io/projected/db7422df-8be7-4a79-9981-b5fd2348e4c4-kube-api-access-sb2qs\") pod \"96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:35 crc kubenswrapper[4789]: I1008 14:16:35.889663 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:36 crc kubenswrapper[4789]: I1008 14:16:36.110334 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq"] Oct 08 14:16:36 crc kubenswrapper[4789]: W1008 14:16:36.120281 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb7422df_8be7_4a79_9981_b5fd2348e4c4.slice/crio-a613f7cb7f1042cbfb40a253050077212b0fd3a00da49589c4e271781828516c WatchSource:0}: Error finding container a613f7cb7f1042cbfb40a253050077212b0fd3a00da49589c4e271781828516c: Status 404 returned error can't find the container with id a613f7cb7f1042cbfb40a253050077212b0fd3a00da49589c4e271781828516c Oct 08 14:16:36 crc kubenswrapper[4789]: I1008 14:16:36.707565 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" event={"ID":"db7422df-8be7-4a79-9981-b5fd2348e4c4","Type":"ContainerStarted","Data":"a613f7cb7f1042cbfb40a253050077212b0fd3a00da49589c4e271781828516c"} Oct 08 14:16:37 crc kubenswrapper[4789]: I1008 14:16:37.715685 4789 generic.go:334] "Generic (PLEG): container finished" podID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerID="42348f7e4b88aa01422ba8a78a90a3012e1d007028ac3ecdf32ffed4ab5fabe8" exitCode=0 Oct 08 14:16:37 crc kubenswrapper[4789]: I1008 14:16:37.715793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" event={"ID":"db7422df-8be7-4a79-9981-b5fd2348e4c4","Type":"ContainerDied","Data":"42348f7e4b88aa01422ba8a78a90a3012e1d007028ac3ecdf32ffed4ab5fabe8"} Oct 08 14:16:38 crc kubenswrapper[4789]: I1008 14:16:38.724304 4789 generic.go:334] "Generic (PLEG): container finished" podID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerID="858cfb0f073271d5e32af6cb2186c8ca34011ff0c795afde0410889960554fa4" exitCode=0 Oct 08 14:16:38 crc kubenswrapper[4789]: I1008 14:16:38.724357 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" event={"ID":"db7422df-8be7-4a79-9981-b5fd2348e4c4","Type":"ContainerDied","Data":"858cfb0f073271d5e32af6cb2186c8ca34011ff0c795afde0410889960554fa4"} Oct 08 14:16:39 crc kubenswrapper[4789]: I1008 14:16:39.735696 4789 generic.go:334] "Generic (PLEG): container finished" podID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerID="53c4258a5b2c509cf65898ee873f3aaca2d34ed18505a73693a41dc89e29f80d" exitCode=0 Oct 08 14:16:39 crc kubenswrapper[4789]: I1008 14:16:39.735766 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" event={"ID":"db7422df-8be7-4a79-9981-b5fd2348e4c4","Type":"ContainerDied","Data":"53c4258a5b2c509cf65898ee873f3aaca2d34ed18505a73693a41dc89e29f80d"} Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.038811 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.201564 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-util\") pod \"db7422df-8be7-4a79-9981-b5fd2348e4c4\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.201703 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-bundle\") pod \"db7422df-8be7-4a79-9981-b5fd2348e4c4\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.201805 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb2qs\" (UniqueName: \"kubernetes.io/projected/db7422df-8be7-4a79-9981-b5fd2348e4c4-kube-api-access-sb2qs\") pod \"db7422df-8be7-4a79-9981-b5fd2348e4c4\" (UID: \"db7422df-8be7-4a79-9981-b5fd2348e4c4\") " Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.202912 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-bundle" (OuterVolumeSpecName: "bundle") pod "db7422df-8be7-4a79-9981-b5fd2348e4c4" (UID: "db7422df-8be7-4a79-9981-b5fd2348e4c4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.207979 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7422df-8be7-4a79-9981-b5fd2348e4c4-kube-api-access-sb2qs" (OuterVolumeSpecName: "kube-api-access-sb2qs") pod "db7422df-8be7-4a79-9981-b5fd2348e4c4" (UID: "db7422df-8be7-4a79-9981-b5fd2348e4c4"). InnerVolumeSpecName "kube-api-access-sb2qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.224178 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-util" (OuterVolumeSpecName: "util") pod "db7422df-8be7-4a79-9981-b5fd2348e4c4" (UID: "db7422df-8be7-4a79-9981-b5fd2348e4c4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.303766 4789 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.303823 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb2qs\" (UniqueName: \"kubernetes.io/projected/db7422df-8be7-4a79-9981-b5fd2348e4c4-kube-api-access-sb2qs\") on node \"crc\" DevicePath \"\"" Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.303843 4789 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db7422df-8be7-4a79-9981-b5fd2348e4c4-util\") on node \"crc\" DevicePath \"\"" Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.753914 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" event={"ID":"db7422df-8be7-4a79-9981-b5fd2348e4c4","Type":"ContainerDied","Data":"a613f7cb7f1042cbfb40a253050077212b0fd3a00da49589c4e271781828516c"} Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.754253 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a613f7cb7f1042cbfb40a253050077212b0fd3a00da49589c4e271781828516c" Oct 08 14:16:41 crc kubenswrapper[4789]: I1008 14:16:41.754069 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq" Oct 08 14:16:49 crc kubenswrapper[4789]: I1008 14:16:49.720997 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r"] Oct 08 14:16:49 crc kubenswrapper[4789]: E1008 14:16:49.721758 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerName="pull" Oct 08 14:16:49 crc kubenswrapper[4789]: I1008 14:16:49.721770 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerName="pull" Oct 08 14:16:49 crc kubenswrapper[4789]: E1008 14:16:49.721781 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerName="extract" Oct 08 14:16:49 crc kubenswrapper[4789]: I1008 14:16:49.721787 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerName="extract" Oct 08 14:16:49 crc kubenswrapper[4789]: E1008 14:16:49.721797 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerName="util" Oct 08 14:16:49 crc kubenswrapper[4789]: I1008 14:16:49.721804 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerName="util" Oct 08 14:16:49 crc kubenswrapper[4789]: I1008 14:16:49.721912 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7422df-8be7-4a79-9981-b5fd2348e4c4" containerName="extract" Oct 08 14:16:49 crc kubenswrapper[4789]: I1008 14:16:49.722585 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" Oct 08 14:16:49 crc kubenswrapper[4789]: I1008 14:16:49.726723 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-w2w52" Oct 08 14:16:49 crc kubenswrapper[4789]: I1008 14:16:49.804590 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r"] Oct 08 14:16:49 crc kubenswrapper[4789]: I1008 14:16:49.916253 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w4pl\" (UniqueName: \"kubernetes.io/projected/9eaa64d7-4d19-436b-a954-ffd75b6d6cfd-kube-api-access-8w4pl\") pod \"openstack-operator-controller-operator-5cf8d5fc8d-4524r\" (UID: \"9eaa64d7-4d19-436b-a954-ffd75b6d6cfd\") " pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" Oct 08 14:16:50 crc kubenswrapper[4789]: I1008 14:16:50.017123 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w4pl\" (UniqueName: \"kubernetes.io/projected/9eaa64d7-4d19-436b-a954-ffd75b6d6cfd-kube-api-access-8w4pl\") pod \"openstack-operator-controller-operator-5cf8d5fc8d-4524r\" (UID: \"9eaa64d7-4d19-436b-a954-ffd75b6d6cfd\") " pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" Oct 08 14:16:50 crc kubenswrapper[4789]: I1008 14:16:50.052952 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w4pl\" (UniqueName: \"kubernetes.io/projected/9eaa64d7-4d19-436b-a954-ffd75b6d6cfd-kube-api-access-8w4pl\") pod \"openstack-operator-controller-operator-5cf8d5fc8d-4524r\" (UID: \"9eaa64d7-4d19-436b-a954-ffd75b6d6cfd\") " pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" Oct 08 14:16:50 crc kubenswrapper[4789]: I1008 14:16:50.345796 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" Oct 08 14:16:50 crc kubenswrapper[4789]: I1008 14:16:50.864217 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r"] Oct 08 14:16:50 crc kubenswrapper[4789]: I1008 14:16:50.875753 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 14:16:51 crc kubenswrapper[4789]: I1008 14:16:51.824774 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" event={"ID":"9eaa64d7-4d19-436b-a954-ffd75b6d6cfd","Type":"ContainerStarted","Data":"3e4e10cdffa2e91c12ab7d092a5a3bcfabd4c6324be23de7f36b79607dc813b5"} Oct 08 14:16:55 crc kubenswrapper[4789]: I1008 14:16:55.869329 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" event={"ID":"9eaa64d7-4d19-436b-a954-ffd75b6d6cfd","Type":"ContainerStarted","Data":"15f9c4e55974c5446f43408647c8c31b83ac906232a07f0c44bcc0ed7abb7220"} Oct 08 14:16:56 crc kubenswrapper[4789]: I1008 14:16:56.433523 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:16:56 crc kubenswrapper[4789]: I1008 14:16:56.433603 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:16:56 crc kubenswrapper[4789]: I1008 14:16:56.433661 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:16:56 crc kubenswrapper[4789]: I1008 14:16:56.434573 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"49beaccc4d4e056fc391f28224570c17529721b7b590ae8c4d56a12f0b6f40ae"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:16:56 crc kubenswrapper[4789]: I1008 14:16:56.434651 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://49beaccc4d4e056fc391f28224570c17529721b7b590ae8c4d56a12f0b6f40ae" gracePeriod=600 Oct 08 14:16:56 crc kubenswrapper[4789]: I1008 14:16:56.876951 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="49beaccc4d4e056fc391f28224570c17529721b7b590ae8c4d56a12f0b6f40ae" exitCode=0 Oct 08 14:16:56 crc kubenswrapper[4789]: I1008 14:16:56.877017 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"49beaccc4d4e056fc391f28224570c17529721b7b590ae8c4d56a12f0b6f40ae"} Oct 08 14:16:56 crc kubenswrapper[4789]: I1008 14:16:56.877072 4789 scope.go:117] "RemoveContainer" containerID="8158c6bcd8d574426927fe18b85723afd7f1a9d1d3d0dd0ab1d58a6c4269820d" Oct 08 14:16:57 crc kubenswrapper[4789]: I1008 14:16:57.908793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"b8550892ad865dfdb5faa70bb82d5784fbdac5d69789616846b92c98439e8b06"} Oct 08 14:16:57 crc kubenswrapper[4789]: I1008 14:16:57.915289 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" event={"ID":"9eaa64d7-4d19-436b-a954-ffd75b6d6cfd","Type":"ContainerStarted","Data":"2ddd6c73ade7a7e8e7afbc921c09df7311aa3138e15cc632418fc864fe13de93"} Oct 08 14:16:57 crc kubenswrapper[4789]: I1008 14:16:57.915836 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" Oct 08 14:16:57 crc kubenswrapper[4789]: I1008 14:16:57.957901 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" podStartSLOduration=2.543137443 podStartE2EDuration="8.957882483s" podCreationTimestamp="2025-10-08 14:16:49 +0000 UTC" firstStartedPulling="2025-10-08 14:16:50.87520597 +0000 UTC m=+950.781953472" lastFinishedPulling="2025-10-08 14:16:57.28995099 +0000 UTC m=+957.196698512" observedRunningTime="2025-10-08 14:16:57.952895067 +0000 UTC m=+957.859642569" watchObservedRunningTime="2025-10-08 14:16:57.957882483 +0000 UTC m=+957.864629975" Oct 08 14:17:00 crc kubenswrapper[4789]: I1008 14:17:00.349833 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5cf8d5fc8d-4524r" Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.946077 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5"] Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.947942 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.952693 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-vxvgn" Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.952852 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt"] Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.953812 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.957781 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-6mdwn" Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.961557 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5"] Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.989760 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsxkj\" (UniqueName: \"kubernetes.io/projected/78f0cb36-5bad-4021-b464-c5da9318404a-kube-api-access-gsxkj\") pod \"cinder-operator-controller-manager-59cdc64769-k8txt\" (UID: \"78f0cb36-5bad-4021-b464-c5da9318404a\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.989817 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7km6n\" (UniqueName: \"kubernetes.io/projected/20bc6d13-f85d-423a-9782-097696fe3a0a-kube-api-access-7km6n\") pod \"barbican-operator-controller-manager-64f84fcdbb-77hn5\" (UID: \"20bc6d13-f85d-423a-9782-097696fe3a0a\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" Oct 08 14:17:32 crc kubenswrapper[4789]: I1008 14:17:32.992537 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.010065 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.011098 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.015138 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-wwnxb" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.031146 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.032496 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.036895 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.044067 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-7glgc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.044765 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.062028 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.074060 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.077562 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-fns6n" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.079592 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.090703 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsxkj\" (UniqueName: \"kubernetes.io/projected/78f0cb36-5bad-4021-b464-c5da9318404a-kube-api-access-gsxkj\") pod \"cinder-operator-controller-manager-59cdc64769-k8txt\" (UID: \"78f0cb36-5bad-4021-b464-c5da9318404a\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.090770 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7km6n\" (UniqueName: \"kubernetes.io/projected/20bc6d13-f85d-423a-9782-097696fe3a0a-kube-api-access-7km6n\") pod \"barbican-operator-controller-manager-64f84fcdbb-77hn5\" (UID: \"20bc6d13-f85d-423a-9782-097696fe3a0a\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.091625 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.092721 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.098786 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-f2t8j" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.106470 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.122920 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7km6n\" (UniqueName: \"kubernetes.io/projected/20bc6d13-f85d-423a-9782-097696fe3a0a-kube-api-access-7km6n\") pod \"barbican-operator-controller-manager-64f84fcdbb-77hn5\" (UID: \"20bc6d13-f85d-423a-9782-097696fe3a0a\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.134300 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsxkj\" (UniqueName: \"kubernetes.io/projected/78f0cb36-5bad-4021-b464-c5da9318404a-kube-api-access-gsxkj\") pod \"cinder-operator-controller-manager-59cdc64769-k8txt\" (UID: \"78f0cb36-5bad-4021-b464-c5da9318404a\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.176431 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.177643 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.180499 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-qw52d" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.180695 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.180849 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.181958 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.186783 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.190686 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-mxfq6" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.196965 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlmr5\" (UniqueName: \"kubernetes.io/projected/0d8d4170-6f10-4d04-b443-297aada2dc10-kube-api-access-hlmr5\") pod \"heat-operator-controller-manager-6d9967f8dd-vbtkd\" (UID: \"0d8d4170-6f10-4d04-b443-297aada2dc10\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.197019 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb89n\" (UniqueName: \"kubernetes.io/projected/e8242a68-738b-4fcd-abaa-22a18790ea28-kube-api-access-tb89n\") pod \"infra-operator-controller-manager-585fc5b659-wl7d2\" (UID: \"e8242a68-738b-4fcd-abaa-22a18790ea28\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.197050 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcvfr\" (UniqueName: \"kubernetes.io/projected/00c51d73-7d74-484a-8773-41cbb40c69f9-kube-api-access-wcvfr\") pod \"glance-operator-controller-manager-7bb46cd7d-tggcc\" (UID: \"00c51d73-7d74-484a-8773-41cbb40c69f9\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.197136 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pqwz\" (UniqueName: \"kubernetes.io/projected/bff33916-174d-4fd4-a5e4-91f7cacb6e52-kube-api-access-4pqwz\") pod \"ironic-operator-controller-manager-74cb5cbc49-zslsw\" (UID: \"bff33916-174d-4fd4-a5e4-91f7cacb6e52\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.197171 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8242a68-738b-4fcd-abaa-22a18790ea28-cert\") pod \"infra-operator-controller-manager-585fc5b659-wl7d2\" (UID: \"e8242a68-738b-4fcd-abaa-22a18790ea28\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.197197 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njtg5\" (UniqueName: \"kubernetes.io/projected/f46571c0-32a7-4202-b308-54446adfe3d1-kube-api-access-njtg5\") pod \"horizon-operator-controller-manager-6d74794d9b-w5bgc\" (UID: \"f46571c0-32a7-4202-b308-54446adfe3d1\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.197284 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfr84\" (UniqueName: \"kubernetes.io/projected/0aa40208-3985-4820-809f-ca1f94b05036-kube-api-access-mfr84\") pod \"designate-operator-controller-manager-687df44cdb-ptjbw\" (UID: \"0aa40208-3985-4820-809f-ca1f94b05036\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.199705 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.203519 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.204738 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.209513 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-m49jx" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.213848 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.214924 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.216893 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-2lf68" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.248675 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.262044 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.263273 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.270347 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-ltj55" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.303832 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.309187 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.310711 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.317979 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8242a68-738b-4fcd-abaa-22a18790ea28-cert\") pod \"infra-operator-controller-manager-585fc5b659-wl7d2\" (UID: \"e8242a68-738b-4fcd-abaa-22a18790ea28\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.318317 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njtg5\" (UniqueName: \"kubernetes.io/projected/f46571c0-32a7-4202-b308-54446adfe3d1-kube-api-access-njtg5\") pod \"horizon-operator-controller-manager-6d74794d9b-w5bgc\" (UID: \"f46571c0-32a7-4202-b308-54446adfe3d1\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.318358 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfr84\" (UniqueName: \"kubernetes.io/projected/0aa40208-3985-4820-809f-ca1f94b05036-kube-api-access-mfr84\") pod \"designate-operator-controller-manager-687df44cdb-ptjbw\" (UID: \"0aa40208-3985-4820-809f-ca1f94b05036\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.318379 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb89n\" (UniqueName: \"kubernetes.io/projected/e8242a68-738b-4fcd-abaa-22a18790ea28-kube-api-access-tb89n\") pod \"infra-operator-controller-manager-585fc5b659-wl7d2\" (UID: \"e8242a68-738b-4fcd-abaa-22a18790ea28\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.318400 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlmr5\" (UniqueName: \"kubernetes.io/projected/0d8d4170-6f10-4d04-b443-297aada2dc10-kube-api-access-hlmr5\") pod \"heat-operator-controller-manager-6d9967f8dd-vbtkd\" (UID: \"0d8d4170-6f10-4d04-b443-297aada2dc10\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.318423 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcvfr\" (UniqueName: \"kubernetes.io/projected/00c51d73-7d74-484a-8773-41cbb40c69f9-kube-api-access-wcvfr\") pod \"glance-operator-controller-manager-7bb46cd7d-tggcc\" (UID: \"00c51d73-7d74-484a-8773-41cbb40c69f9\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.318492 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pqwz\" (UniqueName: \"kubernetes.io/projected/bff33916-174d-4fd4-a5e4-91f7cacb6e52-kube-api-access-4pqwz\") pod \"ironic-operator-controller-manager-74cb5cbc49-zslsw\" (UID: \"bff33916-174d-4fd4-a5e4-91f7cacb6e52\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" Oct 08 14:17:33 crc kubenswrapper[4789]: E1008 14:17:33.318897 4789 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 08 14:17:33 crc kubenswrapper[4789]: E1008 14:17:33.319054 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8242a68-738b-4fcd-abaa-22a18790ea28-cert podName:e8242a68-738b-4fcd-abaa-22a18790ea28 nodeName:}" failed. No retries permitted until 2025-10-08 14:17:33.818976864 +0000 UTC m=+993.725724346 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e8242a68-738b-4fcd-abaa-22a18790ea28-cert") pod "infra-operator-controller-manager-585fc5b659-wl7d2" (UID: "e8242a68-738b-4fcd-abaa-22a18790ea28") : secret "infra-operator-webhook-server-cert" not found Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.353636 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pqwz\" (UniqueName: \"kubernetes.io/projected/bff33916-174d-4fd4-a5e4-91f7cacb6e52-kube-api-access-4pqwz\") pod \"ironic-operator-controller-manager-74cb5cbc49-zslsw\" (UID: \"bff33916-174d-4fd4-a5e4-91f7cacb6e52\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.367651 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcvfr\" (UniqueName: \"kubernetes.io/projected/00c51d73-7d74-484a-8773-41cbb40c69f9-kube-api-access-wcvfr\") pod \"glance-operator-controller-manager-7bb46cd7d-tggcc\" (UID: \"00c51d73-7d74-484a-8773-41cbb40c69f9\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.377798 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb89n\" (UniqueName: \"kubernetes.io/projected/e8242a68-738b-4fcd-abaa-22a18790ea28-kube-api-access-tb89n\") pod \"infra-operator-controller-manager-585fc5b659-wl7d2\" (UID: \"e8242a68-738b-4fcd-abaa-22a18790ea28\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.407100 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.412405 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njtg5\" (UniqueName: \"kubernetes.io/projected/f46571c0-32a7-4202-b308-54446adfe3d1-kube-api-access-njtg5\") pod \"horizon-operator-controller-manager-6d74794d9b-w5bgc\" (UID: \"f46571c0-32a7-4202-b308-54446adfe3d1\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.412821 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlmr5\" (UniqueName: \"kubernetes.io/projected/0d8d4170-6f10-4d04-b443-297aada2dc10-kube-api-access-hlmr5\") pod \"heat-operator-controller-manager-6d9967f8dd-vbtkd\" (UID: \"0d8d4170-6f10-4d04-b443-297aada2dc10\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.414388 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.423872 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.425209 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8p2x\" (UniqueName: \"kubernetes.io/projected/f7b65ead-d9d5-4f06-80de-ae5c85dcbd45-kube-api-access-l8p2x\") pod \"keystone-operator-controller-manager-ddb98f99b-bvjfl\" (UID: \"f7b65ead-d9d5-4f06-80de-ae5c85dcbd45\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.425247 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtd5l\" (UniqueName: \"kubernetes.io/projected/3c0f54fe-d45f-44c2-9f09-4065c9621903-kube-api-access-xtd5l\") pod \"mariadb-operator-controller-manager-5777b4f897-7gsbz\" (UID: \"3c0f54fe-d45f-44c2-9f09-4065c9621903\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.425268 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x9gl\" (UniqueName: \"kubernetes.io/projected/8e51126b-8f41-4627-9f68-fb1255bda5bd-kube-api-access-8x9gl\") pod \"manila-operator-controller-manager-59578bc799-pzg7c\" (UID: \"8e51126b-8f41-4627-9f68-fb1255bda5bd\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.425430 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.428937 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.435388 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-m2lmh" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.439275 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfr84\" (UniqueName: \"kubernetes.io/projected/0aa40208-3985-4820-809f-ca1f94b05036-kube-api-access-mfr84\") pod \"designate-operator-controller-manager-687df44cdb-ptjbw\" (UID: \"0aa40208-3985-4820-809f-ca1f94b05036\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.449079 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.450173 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.454625 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-m9mnj" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.511155 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.529153 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s29rx\" (UniqueName: \"kubernetes.io/projected/c9e3e3c8-447f-4880-b54b-1d5601271216-kube-api-access-s29rx\") pod \"nova-operator-controller-manager-57bb74c7bf-7trlm\" (UID: \"c9e3e3c8-447f-4880-b54b-1d5601271216\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.529244 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8p2x\" (UniqueName: \"kubernetes.io/projected/f7b65ead-d9d5-4f06-80de-ae5c85dcbd45-kube-api-access-l8p2x\") pod \"keystone-operator-controller-manager-ddb98f99b-bvjfl\" (UID: \"f7b65ead-d9d5-4f06-80de-ae5c85dcbd45\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.529279 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8p2k\" (UniqueName: \"kubernetes.io/projected/c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4-kube-api-access-t8p2k\") pod \"neutron-operator-controller-manager-797d478b46-2j5bd\" (UID: \"c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.529298 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtd5l\" (UniqueName: \"kubernetes.io/projected/3c0f54fe-d45f-44c2-9f09-4065c9621903-kube-api-access-xtd5l\") pod \"mariadb-operator-controller-manager-5777b4f897-7gsbz\" (UID: \"3c0f54fe-d45f-44c2-9f09-4065c9621903\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.529317 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x9gl\" (UniqueName: \"kubernetes.io/projected/8e51126b-8f41-4627-9f68-fb1255bda5bd-kube-api-access-8x9gl\") pod \"manila-operator-controller-manager-59578bc799-pzg7c\" (UID: \"8e51126b-8f41-4627-9f68-fb1255bda5bd\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.551000 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtd5l\" (UniqueName: \"kubernetes.io/projected/3c0f54fe-d45f-44c2-9f09-4065c9621903-kube-api-access-xtd5l\") pod \"mariadb-operator-controller-manager-5777b4f897-7gsbz\" (UID: \"3c0f54fe-d45f-44c2-9f09-4065c9621903\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.554393 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.558165 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.564600 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8p2x\" (UniqueName: \"kubernetes.io/projected/f7b65ead-d9d5-4f06-80de-ae5c85dcbd45-kube-api-access-l8p2x\") pod \"keystone-operator-controller-manager-ddb98f99b-bvjfl\" (UID: \"f7b65ead-d9d5-4f06-80de-ae5c85dcbd45\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.586530 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x9gl\" (UniqueName: \"kubernetes.io/projected/8e51126b-8f41-4627-9f68-fb1255bda5bd-kube-api-access-8x9gl\") pod \"manila-operator-controller-manager-59578bc799-pzg7c\" (UID: \"8e51126b-8f41-4627-9f68-fb1255bda5bd\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.602020 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.608053 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.609202 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.611862 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-79jhl" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.632588 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s29rx\" (UniqueName: \"kubernetes.io/projected/c9e3e3c8-447f-4880-b54b-1d5601271216-kube-api-access-s29rx\") pod \"nova-operator-controller-manager-57bb74c7bf-7trlm\" (UID: \"c9e3e3c8-447f-4880-b54b-1d5601271216\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.632662 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8p2k\" (UniqueName: \"kubernetes.io/projected/c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4-kube-api-access-t8p2k\") pod \"neutron-operator-controller-manager-797d478b46-2j5bd\" (UID: \"c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.635763 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.636743 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.649178 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.650566 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.650830 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.651170 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.652058 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.654292 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.657532 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.657644 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-jz55t" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.657540 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-nhgm8" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.677634 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.686541 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8p2k\" (UniqueName: \"kubernetes.io/projected/c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4-kube-api-access-t8p2k\") pod \"neutron-operator-controller-manager-797d478b46-2j5bd\" (UID: \"c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.689948 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s29rx\" (UniqueName: \"kubernetes.io/projected/c9e3e3c8-447f-4880-b54b-1d5601271216-kube-api-access-s29rx\") pod \"nova-operator-controller-manager-57bb74c7bf-7trlm\" (UID: \"c9e3e3c8-447f-4880-b54b-1d5601271216\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.697387 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.698785 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.705448 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.705959 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-7qwcm" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.723143 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.724334 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.730096 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-f4cb5" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.735021 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.735848 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtfwm\" (UniqueName: \"kubernetes.io/projected/4e7cd567-305f-417d-bac8-7bf4caf22693-kube-api-access-wtfwm\") pod \"octavia-operator-controller-manager-6d7c7ddf95-2kzff\" (UID: \"4e7cd567-305f-417d-bac8-7bf4caf22693\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.736115 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.741064 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-4gknm" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.747204 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.748728 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.759863 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.770044 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.771571 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.785083 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-n4vtd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.785260 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.796128 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.797341 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.805278 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-p5wcv" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.820555 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.820866 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.837106 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjzlh\" (UniqueName: \"kubernetes.io/projected/67d8c63c-cc55-4be0-bea4-8371ac8e333b-kube-api-access-mjzlh\") pod \"placement-operator-controller-manager-664664cb68-4xgrb\" (UID: \"67d8c63c-cc55-4be0-bea4-8371ac8e333b\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.837151 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9ndj\" (UniqueName: \"kubernetes.io/projected/5578f70a-ad1b-47f8-a6ce-5dc726788ded-kube-api-access-r9ndj\") pod \"ovn-operator-controller-manager-6f96f8c84-6f9x9\" (UID: \"5578f70a-ad1b-47f8-a6ce-5dc726788ded\") " pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.837210 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52xzt\" (UniqueName: \"kubernetes.io/projected/bc485c37-780a-4509-b3f3-aaf1714a9edb-kube-api-access-52xzt\") pod \"swift-operator-controller-manager-5f4d5dfdc6-nfllk\" (UID: \"bc485c37-780a-4509-b3f3-aaf1714a9edb\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.837251 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8242a68-738b-4fcd-abaa-22a18790ea28-cert\") pod \"infra-operator-controller-manager-585fc5b659-wl7d2\" (UID: \"e8242a68-738b-4fcd-abaa-22a18790ea28\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.837268 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f893fd6-e5cc-42b9-ac46-77567818c0ef-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h\" (UID: \"6f893fd6-e5cc-42b9-ac46-77567818c0ef\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.837296 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mw6b\" (UniqueName: \"kubernetes.io/projected/6f893fd6-e5cc-42b9-ac46-77567818c0ef-kube-api-access-6mw6b\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h\" (UID: \"6f893fd6-e5cc-42b9-ac46-77567818c0ef\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.837317 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72j2r\" (UniqueName: \"kubernetes.io/projected/8efa82a4-efc9-4abe-b373-8828def5bb61-kube-api-access-72j2r\") pod \"telemetry-operator-controller-manager-775776c574-mgjfp\" (UID: \"8efa82a4-efc9-4abe-b373-8828def5bb61\") " pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.837355 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtfwm\" (UniqueName: \"kubernetes.io/projected/4e7cd567-305f-417d-bac8-7bf4caf22693-kube-api-access-wtfwm\") pod \"octavia-operator-controller-manager-6d7c7ddf95-2kzff\" (UID: \"4e7cd567-305f-417d-bac8-7bf4caf22693\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" Oct 08 14:17:33 crc kubenswrapper[4789]: E1008 14:17:33.837703 4789 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 08 14:17:33 crc kubenswrapper[4789]: E1008 14:17:33.837747 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8242a68-738b-4fcd-abaa-22a18790ea28-cert podName:e8242a68-738b-4fcd-abaa-22a18790ea28 nodeName:}" failed. No retries permitted until 2025-10-08 14:17:34.837732023 +0000 UTC m=+994.744479515 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e8242a68-738b-4fcd-abaa-22a18790ea28-cert") pod "infra-operator-controller-manager-585fc5b659-wl7d2" (UID: "e8242a68-738b-4fcd-abaa-22a18790ea28") : secret "infra-operator-webhook-server-cert" not found Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.853722 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.868313 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.897957 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtfwm\" (UniqueName: \"kubernetes.io/projected/4e7cd567-305f-417d-bac8-7bf4caf22693-kube-api-access-wtfwm\") pod \"octavia-operator-controller-manager-6d7c7ddf95-2kzff\" (UID: \"4e7cd567-305f-417d-bac8-7bf4caf22693\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.910336 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.911706 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.917949 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.918195 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-tpxcf" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942049 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjzlh\" (UniqueName: \"kubernetes.io/projected/67d8c63c-cc55-4be0-bea4-8371ac8e333b-kube-api-access-mjzlh\") pod \"placement-operator-controller-manager-664664cb68-4xgrb\" (UID: \"67d8c63c-cc55-4be0-bea4-8371ac8e333b\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942129 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9ndj\" (UniqueName: \"kubernetes.io/projected/5578f70a-ad1b-47f8-a6ce-5dc726788ded-kube-api-access-r9ndj\") pod \"ovn-operator-controller-manager-6f96f8c84-6f9x9\" (UID: \"5578f70a-ad1b-47f8-a6ce-5dc726788ded\") " pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942199 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52xzt\" (UniqueName: \"kubernetes.io/projected/bc485c37-780a-4509-b3f3-aaf1714a9edb-kube-api-access-52xzt\") pod \"swift-operator-controller-manager-5f4d5dfdc6-nfllk\" (UID: \"bc485c37-780a-4509-b3f3-aaf1714a9edb\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942243 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkhq6\" (UniqueName: \"kubernetes.io/projected/8bdbf6d1-353c-42ba-80df-e60d6b424f55-kube-api-access-jkhq6\") pod \"watcher-operator-controller-manager-6c7dbfbcbd-7xrj5\" (UID: \"8bdbf6d1-353c-42ba-80df-e60d6b424f55\") " pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942277 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d174fba3-84a1-4de8-8b32-4f97d1880a78-cert\") pod \"openstack-operator-controller-manager-766b688d6c-zfshx\" (UID: \"d174fba3-84a1-4de8-8b32-4f97d1880a78\") " pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942330 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f893fd6-e5cc-42b9-ac46-77567818c0ef-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h\" (UID: \"6f893fd6-e5cc-42b9-ac46-77567818c0ef\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942376 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mw6b\" (UniqueName: \"kubernetes.io/projected/6f893fd6-e5cc-42b9-ac46-77567818c0ef-kube-api-access-6mw6b\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h\" (UID: \"6f893fd6-e5cc-42b9-ac46-77567818c0ef\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942413 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72j2r\" (UniqueName: \"kubernetes.io/projected/8efa82a4-efc9-4abe-b373-8828def5bb61-kube-api-access-72j2r\") pod \"telemetry-operator-controller-manager-775776c574-mgjfp\" (UID: \"8efa82a4-efc9-4abe-b373-8828def5bb61\") " pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942447 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7vtn\" (UniqueName: \"kubernetes.io/projected/d174fba3-84a1-4de8-8b32-4f97d1880a78-kube-api-access-n7vtn\") pod \"openstack-operator-controller-manager-766b688d6c-zfshx\" (UID: \"d174fba3-84a1-4de8-8b32-4f97d1880a78\") " pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.942486 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm5jw\" (UniqueName: \"kubernetes.io/projected/08eb1945-dc5b-4f43-9fad-cd2b7e3102be-kube-api-access-vm5jw\") pod \"test-operator-controller-manager-74665f6cdc-x52cc\" (UID: \"08eb1945-dc5b-4f43-9fad-cd2b7e3102be\") " pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" Oct 08 14:17:33 crc kubenswrapper[4789]: E1008 14:17:33.943261 4789 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 14:17:33 crc kubenswrapper[4789]: E1008 14:17:33.943338 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f893fd6-e5cc-42b9-ac46-77567818c0ef-cert podName:6f893fd6-e5cc-42b9-ac46-77567818c0ef nodeName:}" failed. No retries permitted until 2025-10-08 14:17:34.443317154 +0000 UTC m=+994.350064646 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6f893fd6-e5cc-42b9-ac46-77567818c0ef-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" (UID: "6f893fd6-e5cc-42b9-ac46-77567818c0ef") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.953311 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx"] Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.956914 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.977559 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9ndj\" (UniqueName: \"kubernetes.io/projected/5578f70a-ad1b-47f8-a6ce-5dc726788ded-kube-api-access-r9ndj\") pod \"ovn-operator-controller-manager-6f96f8c84-6f9x9\" (UID: \"5578f70a-ad1b-47f8-a6ce-5dc726788ded\") " pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.979602 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52xzt\" (UniqueName: \"kubernetes.io/projected/bc485c37-780a-4509-b3f3-aaf1714a9edb-kube-api-access-52xzt\") pod \"swift-operator-controller-manager-5f4d5dfdc6-nfllk\" (UID: \"bc485c37-780a-4509-b3f3-aaf1714a9edb\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.983618 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mw6b\" (UniqueName: \"kubernetes.io/projected/6f893fd6-e5cc-42b9-ac46-77567818c0ef-kube-api-access-6mw6b\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h\" (UID: \"6f893fd6-e5cc-42b9-ac46-77567818c0ef\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.983896 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72j2r\" (UniqueName: \"kubernetes.io/projected/8efa82a4-efc9-4abe-b373-8828def5bb61-kube-api-access-72j2r\") pod \"telemetry-operator-controller-manager-775776c574-mgjfp\" (UID: \"8efa82a4-efc9-4abe-b373-8828def5bb61\") " pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" Oct 08 14:17:33 crc kubenswrapper[4789]: I1008 14:17:33.996957 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.000962 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.004096 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjzlh\" (UniqueName: \"kubernetes.io/projected/67d8c63c-cc55-4be0-bea4-8371ac8e333b-kube-api-access-mjzlh\") pod \"placement-operator-controller-manager-664664cb68-4xgrb\" (UID: \"67d8c63c-cc55-4be0-bea4-8371ac8e333b\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.022701 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.027938 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.031306 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.034935 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-9mw2z" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.044110 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm5jw\" (UniqueName: \"kubernetes.io/projected/08eb1945-dc5b-4f43-9fad-cd2b7e3102be-kube-api-access-vm5jw\") pod \"test-operator-controller-manager-74665f6cdc-x52cc\" (UID: \"08eb1945-dc5b-4f43-9fad-cd2b7e3102be\") " pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.044199 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkhq6\" (UniqueName: \"kubernetes.io/projected/8bdbf6d1-353c-42ba-80df-e60d6b424f55-kube-api-access-jkhq6\") pod \"watcher-operator-controller-manager-6c7dbfbcbd-7xrj5\" (UID: \"8bdbf6d1-353c-42ba-80df-e60d6b424f55\") " pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.044224 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d174fba3-84a1-4de8-8b32-4f97d1880a78-cert\") pod \"openstack-operator-controller-manager-766b688d6c-zfshx\" (UID: \"d174fba3-84a1-4de8-8b32-4f97d1880a78\") " pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.044457 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7vtn\" (UniqueName: \"kubernetes.io/projected/d174fba3-84a1-4de8-8b32-4f97d1880a78-kube-api-access-n7vtn\") pod \"openstack-operator-controller-manager-766b688d6c-zfshx\" (UID: \"d174fba3-84a1-4de8-8b32-4f97d1880a78\") " pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:34 crc kubenswrapper[4789]: E1008 14:17:34.044972 4789 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 08 14:17:34 crc kubenswrapper[4789]: E1008 14:17:34.045307 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d174fba3-84a1-4de8-8b32-4f97d1880a78-cert podName:d174fba3-84a1-4de8-8b32-4f97d1880a78 nodeName:}" failed. No retries permitted until 2025-10-08 14:17:34.545291128 +0000 UTC m=+994.452038620 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d174fba3-84a1-4de8-8b32-4f97d1880a78-cert") pod "openstack-operator-controller-manager-766b688d6c-zfshx" (UID: "d174fba3-84a1-4de8-8b32-4f97d1880a78") : secret "webhook-server-cert" not found Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.062336 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.072975 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm5jw\" (UniqueName: \"kubernetes.io/projected/08eb1945-dc5b-4f43-9fad-cd2b7e3102be-kube-api-access-vm5jw\") pod \"test-operator-controller-manager-74665f6cdc-x52cc\" (UID: \"08eb1945-dc5b-4f43-9fad-cd2b7e3102be\") " pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.082689 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7vtn\" (UniqueName: \"kubernetes.io/projected/d174fba3-84a1-4de8-8b32-4f97d1880a78-kube-api-access-n7vtn\") pod \"openstack-operator-controller-manager-766b688d6c-zfshx\" (UID: \"d174fba3-84a1-4de8-8b32-4f97d1880a78\") " pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.086853 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkhq6\" (UniqueName: \"kubernetes.io/projected/8bdbf6d1-353c-42ba-80df-e60d6b424f55-kube-api-access-jkhq6\") pod \"watcher-operator-controller-manager-6c7dbfbcbd-7xrj5\" (UID: \"8bdbf6d1-353c-42ba-80df-e60d6b424f55\") " pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.146054 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5fs9\" (UniqueName: \"kubernetes.io/projected/8bc03843-9303-451e-b933-c450aaad7401-kube-api-access-n5fs9\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5\" (UID: \"8bc03843-9303-451e-b933-c450aaad7401\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.247807 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5fs9\" (UniqueName: \"kubernetes.io/projected/8bc03843-9303-451e-b933-c450aaad7401-kube-api-access-n5fs9\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5\" (UID: \"8bc03843-9303-451e-b933-c450aaad7401\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.272878 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5fs9\" (UniqueName: \"kubernetes.io/projected/8bc03843-9303-451e-b933-c450aaad7401-kube-api-access-n5fs9\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5\" (UID: \"8bc03843-9303-451e-b933-c450aaad7401\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.279194 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.311059 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.337867 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.452687 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f893fd6-e5cc-42b9-ac46-77567818c0ef-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h\" (UID: \"6f893fd6-e5cc-42b9-ac46-77567818c0ef\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.458738 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6f893fd6-e5cc-42b9-ac46-77567818c0ef-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h\" (UID: \"6f893fd6-e5cc-42b9-ac46-77567818c0ef\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.474475 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.518713 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.534375 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.553797 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d174fba3-84a1-4de8-8b32-4f97d1880a78-cert\") pod \"openstack-operator-controller-manager-766b688d6c-zfshx\" (UID: \"d174fba3-84a1-4de8-8b32-4f97d1880a78\") " pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.560359 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d174fba3-84a1-4de8-8b32-4f97d1880a78-cert\") pod \"openstack-operator-controller-manager-766b688d6c-zfshx\" (UID: \"d174fba3-84a1-4de8-8b32-4f97d1880a78\") " pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.582140 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.584261 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.677388 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.781235 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.785668 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5"] Oct 08 14:17:34 crc kubenswrapper[4789]: W1008 14:17:34.809099 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbff33916_174d_4fd4_a5e4_91f7cacb6e52.slice/crio-9f7d4847cb9fbcbb632cec205a71958a9763b9ae93a8c96604c92b00cd773a29 WatchSource:0}: Error finding container 9f7d4847cb9fbcbb632cec205a71958a9763b9ae93a8c96604c92b00cd773a29: Status 404 returned error can't find the container with id 9f7d4847cb9fbcbb632cec205a71958a9763b9ae93a8c96604c92b00cd773a29 Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.858295 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8242a68-738b-4fcd-abaa-22a18790ea28-cert\") pod \"infra-operator-controller-manager-585fc5b659-wl7d2\" (UID: \"e8242a68-738b-4fcd-abaa-22a18790ea28\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.870955 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8242a68-738b-4fcd-abaa-22a18790ea28-cert\") pod \"infra-operator-controller-manager-585fc5b659-wl7d2\" (UID: \"e8242a68-738b-4fcd-abaa-22a18790ea28\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.948904 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.966349 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.970651 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.980254 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl"] Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.988034 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c"] Oct 08 14:17:34 crc kubenswrapper[4789]: W1008 14:17:34.992057 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e51126b_8f41_4627_9f68_fb1255bda5bd.slice/crio-0d293a7a669991251827b5bf489f5e8e05f4030fdc8bdda7bddefd2542689894 WatchSource:0}: Error finding container 0d293a7a669991251827b5bf489f5e8e05f4030fdc8bdda7bddefd2542689894: Status 404 returned error can't find the container with id 0d293a7a669991251827b5bf489f5e8e05f4030fdc8bdda7bddefd2542689894 Oct 08 14:17:34 crc kubenswrapper[4789]: I1008 14:17:34.992818 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw"] Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.018066 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.175145 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" event={"ID":"0d8d4170-6f10-4d04-b443-297aada2dc10","Type":"ContainerStarted","Data":"dcb593a2ba32bad265e60ed70d043862785c461876b13ab30254386a20d60ddd"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.176032 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" event={"ID":"20bc6d13-f85d-423a-9782-097696fe3a0a","Type":"ContainerStarted","Data":"34fcd306f4e6fcee31ac0af63e987665c2eafc4d86174d6f02e6114e9d65c151"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.176632 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" event={"ID":"78f0cb36-5bad-4021-b464-c5da9318404a","Type":"ContainerStarted","Data":"758c393dd36c66b908fbd9ba970c710a518ccbd6ec4a4646b4bfb5904f6da887"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.177254 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" event={"ID":"c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4","Type":"ContainerStarted","Data":"167793db73d77f93ad1822a0e58b371ec479ff1d57c9d5ecb3651fd6e43a1342"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.177856 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" event={"ID":"0aa40208-3985-4820-809f-ca1f94b05036","Type":"ContainerStarted","Data":"ad2b7b7e2fec9797a5bd77d7e2a863fa49c6745195cb8a7fd2370d7c99e8cf6a"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.181873 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" event={"ID":"bff33916-174d-4fd4-a5e4-91f7cacb6e52","Type":"ContainerStarted","Data":"9f7d4847cb9fbcbb632cec205a71958a9763b9ae93a8c96604c92b00cd773a29"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.185006 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" event={"ID":"8e51126b-8f41-4627-9f68-fb1255bda5bd","Type":"ContainerStarted","Data":"0d293a7a669991251827b5bf489f5e8e05f4030fdc8bdda7bddefd2542689894"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.186834 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" event={"ID":"f46571c0-32a7-4202-b308-54446adfe3d1","Type":"ContainerStarted","Data":"54373f82ad5355084203210cbf55d949451207045c69d22d9d3a2fb800e57a36"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.187878 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" event={"ID":"f7b65ead-d9d5-4f06-80de-ae5c85dcbd45","Type":"ContainerStarted","Data":"f25a765092db91e5df87c4b6c0cf8145370d17fa93d93f745f10add7c196fae0"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.189481 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" event={"ID":"c9e3e3c8-447f-4880-b54b-1d5601271216","Type":"ContainerStarted","Data":"39edf3ab9feae0963232f31084b12603d78831b0abc1c562ad4a1c45118da526"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.193558 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" event={"ID":"5578f70a-ad1b-47f8-a6ce-5dc726788ded","Type":"ContainerStarted","Data":"bff620ebc5b854e991ccaf46a4eab92278bcea287e947b102aacb2cbd884a8c5"} Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.390907 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz"] Oct 08 14:17:35 crc kubenswrapper[4789]: W1008 14:17:35.393897 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bc03843_9303_451e_b933_c450aaad7401.slice/crio-04b344a7a93fdf5120ea3066e29267a87c9e9564c6f0d6a061b9ace44aee31e6 WatchSource:0}: Error finding container 04b344a7a93fdf5120ea3066e29267a87c9e9564c6f0d6a061b9ace44aee31e6: Status 404 returned error can't find the container with id 04b344a7a93fdf5120ea3066e29267a87c9e9564c6f0d6a061b9ace44aee31e6 Oct 08 14:17:35 crc kubenswrapper[4789]: W1008 14:17:35.403491 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08eb1945_dc5b_4f43_9fad_cd2b7e3102be.slice/crio-049e7e76dbafbd03013d995dfb07145897806aeb81de4843b1f11ecf88c64f7c WatchSource:0}: Error finding container 049e7e76dbafbd03013d995dfb07145897806aeb81de4843b1f11ecf88c64f7c: Status 404 returned error can't find the container with id 049e7e76dbafbd03013d995dfb07145897806aeb81de4843b1f11ecf88c64f7c Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.403878 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc"] Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.412630 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5"] Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.417134 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk"] Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.436200 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc"] Oct 08 14:17:35 crc kubenswrapper[4789]: E1008 14:17:35.441680 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-52xzt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f4d5dfdc6-nfllk_openstack-operators(bc485c37-780a-4509-b3f3-aaf1714a9edb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.449279 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb"] Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.450564 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx"] Oct 08 14:17:35 crc kubenswrapper[4789]: E1008 14:17:35.465203 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:3cc6bba71197ddf88dd4ba1301542bacbc1fe12e6faab2b69e6960944b3d74a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wcvfr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-7bb46cd7d-tggcc_openstack-operators(00c51d73-7d74-484a-8773-41cbb40c69f9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 14:17:35 crc kubenswrapper[4789]: E1008 14:17:35.466391 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mjzlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-664664cb68-4xgrb_openstack-operators(67d8c63c-cc55-4be0-bea4-8371ac8e333b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 14:17:35 crc kubenswrapper[4789]: E1008 14:17:35.473671 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:9d26476523320d70d6d457b91663e8c233ed320d77032a7c57a89ce1aedd3931,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-72j2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-775776c574-mgjfp_openstack-operators(8efa82a4-efc9-4abe-b373-8828def5bb61): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 14:17:35 crc kubenswrapper[4789]: E1008 14:17:35.477069 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.222:5001/openstack-k8s-operators/watcher-operator:33900f6764443fc3b205c350eb74d5a64f4d6ad4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jkhq6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c7dbfbcbd-7xrj5_openstack-operators(8bdbf6d1-353c-42ba-80df-e60d6b424f55): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.505391 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp"] Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.514828 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5"] Oct 08 14:17:35 crc kubenswrapper[4789]: E1008 14:17:35.520217 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wtfwm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-6d7c7ddf95-2kzff_openstack-operators(4e7cd567-305f-417d-bac8-7bf4caf22693): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 14:17:35 crc kubenswrapper[4789]: E1008 14:17:35.523063 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6mw6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h_openstack-operators(6f893fd6-e5cc-42b9-ac46-77567818c0ef): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.530773 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff"] Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.535705 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h"] Oct 08 14:17:35 crc kubenswrapper[4789]: I1008 14:17:35.539512 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2"] Oct 08 14:17:35 crc kubenswrapper[4789]: E1008 14:17:35.935538 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" podUID="bc485c37-780a-4509-b3f3-aaf1714a9edb" Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.086759 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" podUID="67d8c63c-cc55-4be0-bea4-8371ac8e333b" Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.102418 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" podUID="00c51d73-7d74-484a-8773-41cbb40c69f9" Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.102567 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" podUID="4e7cd567-305f-417d-bac8-7bf4caf22693" Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.223273 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" event={"ID":"67d8c63c-cc55-4be0-bea4-8371ac8e333b","Type":"ContainerStarted","Data":"09413c3a4462c2356d945730d43c0d799a54f77a2c1f2436492976fa1bf05544"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.223660 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" event={"ID":"67d8c63c-cc55-4be0-bea4-8371ac8e333b","Type":"ContainerStarted","Data":"bc1f6a6bdaa782a37cf19a8ad90bb68eed11130edd406847c7a1cd6ab1724e0a"} Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.224588 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" podUID="67d8c63c-cc55-4be0-bea4-8371ac8e333b" Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.225283 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5" event={"ID":"8bc03843-9303-451e-b933-c450aaad7401","Type":"ContainerStarted","Data":"04b344a7a93fdf5120ea3066e29267a87c9e9564c6f0d6a061b9ace44aee31e6"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.232764 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" event={"ID":"00c51d73-7d74-484a-8773-41cbb40c69f9","Type":"ContainerStarted","Data":"0bdf2d3fef9911922d9c46b5d22f62c6cc5c182c956b7b7ec3fd98329e00c59d"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.232806 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" event={"ID":"00c51d73-7d74-484a-8773-41cbb40c69f9","Type":"ContainerStarted","Data":"ed9c0d2e6ceb044b1263b31b232aae7974bbdd11505035f89ba22b7cd3e47eea"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.235774 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" event={"ID":"08eb1945-dc5b-4f43-9fad-cd2b7e3102be","Type":"ContainerStarted","Data":"049e7e76dbafbd03013d995dfb07145897806aeb81de4843b1f11ecf88c64f7c"} Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.235802 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:3cc6bba71197ddf88dd4ba1301542bacbc1fe12e6faab2b69e6960944b3d74a0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" podUID="00c51d73-7d74-484a-8773-41cbb40c69f9" Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.236558 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" event={"ID":"3c0f54fe-d45f-44c2-9f09-4065c9621903","Type":"ContainerStarted","Data":"e954bc2704001ea342072eb081d33db7ce60168694ba49a4b1ce4d588892d632"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.237288 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" event={"ID":"e8242a68-738b-4fcd-abaa-22a18790ea28","Type":"ContainerStarted","Data":"56d0f9feae4e0de9a4c170c60694556356002b4079a705bd82330ed34ae85c62"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.242149 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" event={"ID":"8efa82a4-efc9-4abe-b373-8828def5bb61","Type":"ContainerStarted","Data":"3945658018e957a12833ee3c37c9bfcf23a186ce6539184706d73b9f02a783fd"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.242187 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" event={"ID":"8efa82a4-efc9-4abe-b373-8828def5bb61","Type":"ContainerStarted","Data":"54759a9966ab71ba6421fd3c5cc9c44da08e70aba77015edb5fefd2b94e02583"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.254241 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" event={"ID":"6f893fd6-e5cc-42b9-ac46-77567818c0ef","Type":"ContainerStarted","Data":"bb72776e26708ba507e90b7800b62ab2ee5c2d0a6225ef21969bcc266bb8eb40"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.254299 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" event={"ID":"6f893fd6-e5cc-42b9-ac46-77567818c0ef","Type":"ContainerStarted","Data":"da9e2cb310518fa8a905714b5f313362b5c748414111c028dd2ba10644b52798"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.255638 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" event={"ID":"8bdbf6d1-353c-42ba-80df-e60d6b424f55","Type":"ContainerStarted","Data":"f91ec1369bab8c744c64d457bbc86b2bd88e80f3452e60bb8df326f4896c896e"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.255685 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" event={"ID":"8bdbf6d1-353c-42ba-80df-e60d6b424f55","Type":"ContainerStarted","Data":"5bfbc01ef421b2c83dcb98adf95ace72be565f31f840d5a8cd67b7d6cfd26022"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.256614 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" event={"ID":"d174fba3-84a1-4de8-8b32-4f97d1880a78","Type":"ContainerStarted","Data":"683738658e3bbd8b1b7c2c735a2a81eff43c2abd9bc4452ca520c341c8acb6ba"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.256638 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" event={"ID":"d174fba3-84a1-4de8-8b32-4f97d1880a78","Type":"ContainerStarted","Data":"f41c1ac53e07e1ae9de19ada6592d0ff3dafc57eac6851b843b02c8a54985bca"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.259029 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" event={"ID":"4e7cd567-305f-417d-bac8-7bf4caf22693","Type":"ContainerStarted","Data":"f59e516e1798598274f7d09bd57c1c94bdc1741570af3c3dbb1149598f0d8f88"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.259060 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" event={"ID":"4e7cd567-305f-417d-bac8-7bf4caf22693","Type":"ContainerStarted","Data":"92cf74ab62294e1a3778ecf9de429e413317bd3bdc2668d484bdfe4db02f213c"} Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.261333 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" podUID="4e7cd567-305f-417d-bac8-7bf4caf22693" Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.269245 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" event={"ID":"bc485c37-780a-4509-b3f3-aaf1714a9edb","Type":"ContainerStarted","Data":"08381d4b0e7ef576d9053c9cf6de8afb4d589bec334435ec0238698916cb05d5"} Oct 08 14:17:36 crc kubenswrapper[4789]: I1008 14:17:36.269288 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" event={"ID":"bc485c37-780a-4509-b3f3-aaf1714a9edb","Type":"ContainerStarted","Data":"d39cdd4d9a4b5a71930a1ff305227bc7d58ceb9c3523a380fb17f26efdfcfbf0"} Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.272285 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" podUID="bc485c37-780a-4509-b3f3-aaf1714a9edb" Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.351445 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" podUID="8bdbf6d1-353c-42ba-80df-e60d6b424f55" Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.351622 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" podUID="8efa82a4-efc9-4abe-b373-8828def5bb61" Oct 08 14:17:36 crc kubenswrapper[4789]: E1008 14:17:36.351722 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" podUID="6f893fd6-e5cc-42b9-ac46-77567818c0ef" Oct 08 14:17:37 crc kubenswrapper[4789]: I1008 14:17:37.278088 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" event={"ID":"d174fba3-84a1-4de8-8b32-4f97d1880a78","Type":"ContainerStarted","Data":"9470e3b49c442815a277c7f54f87fe8c87bbb1c0d1300c6fc2a9440fa777a8b3"} Oct 08 14:17:37 crc kubenswrapper[4789]: E1008 14:17:37.281013 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/openstack-k8s-operators/watcher-operator:33900f6764443fc3b205c350eb74d5a64f4d6ad4\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" podUID="8bdbf6d1-353c-42ba-80df-e60d6b424f55" Oct 08 14:17:37 crc kubenswrapper[4789]: E1008 14:17:37.281345 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:9d26476523320d70d6d457b91663e8c233ed320d77032a7c57a89ce1aedd3931\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" podUID="8efa82a4-efc9-4abe-b373-8828def5bb61" Oct 08 14:17:37 crc kubenswrapper[4789]: E1008 14:17:37.281380 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" podUID="6f893fd6-e5cc-42b9-ac46-77567818c0ef" Oct 08 14:17:37 crc kubenswrapper[4789]: E1008 14:17:37.281428 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" podUID="67d8c63c-cc55-4be0-bea4-8371ac8e333b" Oct 08 14:17:37 crc kubenswrapper[4789]: E1008 14:17:37.281485 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" podUID="4e7cd567-305f-417d-bac8-7bf4caf22693" Oct 08 14:17:37 crc kubenswrapper[4789]: E1008 14:17:37.281494 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" podUID="bc485c37-780a-4509-b3f3-aaf1714a9edb" Oct 08 14:17:37 crc kubenswrapper[4789]: E1008 14:17:37.282301 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:3cc6bba71197ddf88dd4ba1301542bacbc1fe12e6faab2b69e6960944b3d74a0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" podUID="00c51d73-7d74-484a-8773-41cbb40c69f9" Oct 08 14:17:37 crc kubenswrapper[4789]: I1008 14:17:37.385997 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" podStartSLOduration=4.385968706 podStartE2EDuration="4.385968706s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:17:37.382943614 +0000 UTC m=+997.289691116" watchObservedRunningTime="2025-10-08 14:17:37.385968706 +0000 UTC m=+997.292716198" Oct 08 14:17:38 crc kubenswrapper[4789]: I1008 14:17:38.283563 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:44 crc kubenswrapper[4789]: I1008 14:17:44.684494 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-766b688d6c-zfshx" Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.406087 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" event={"ID":"08eb1945-dc5b-4f43-9fad-cd2b7e3102be","Type":"ContainerStarted","Data":"8661589d9b4138e8679592f19e5b4d0b7d90e88bc886eadc32234ae85d272efc"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.414946 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" event={"ID":"c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4","Type":"ContainerStarted","Data":"8b977f6fa9a57f22961317d2aa8a0e89a5b27a96d198d388dc3ec0dd0331a1c0"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.422598 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" event={"ID":"c9e3e3c8-447f-4880-b54b-1d5601271216","Type":"ContainerStarted","Data":"0826f110877815cf82e00b6b78551cf88f04a9fb390be371e4a6660b54a9d794"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.426421 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" event={"ID":"bff33916-174d-4fd4-a5e4-91f7cacb6e52","Type":"ContainerStarted","Data":"72bb7729523a8349565b2dfc66e2e042cdf12133950f346087a39a8da3cf5f37"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.450531 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" event={"ID":"5578f70a-ad1b-47f8-a6ce-5dc726788ded","Type":"ContainerStarted","Data":"4c9b187aeda05bc1b41f848766b1739c2e01e4b3cd637a7c16544131854ffdc3"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.453281 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" event={"ID":"0d8d4170-6f10-4d04-b443-297aada2dc10","Type":"ContainerStarted","Data":"6e3310d88af00df41c8ca94c749b7995da0aca66d520b2a083206754dc715a18"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.457617 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" event={"ID":"8e51126b-8f41-4627-9f68-fb1255bda5bd","Type":"ContainerStarted","Data":"51149b22e08c2014fa0e1ef9dc8610a4d371e5a4c52f04d805458f048d0b2254"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.462209 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" event={"ID":"3c0f54fe-d45f-44c2-9f09-4065c9621903","Type":"ContainerStarted","Data":"74313d7cb53380c9c6ff2bb5bb6fed6c0bc9b6199e88c6d4b33440d5da8e87ce"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.467437 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" event={"ID":"20bc6d13-f85d-423a-9782-097696fe3a0a","Type":"ContainerStarted","Data":"e005c9b2f9a2d99c5b47fe4d753bbc82edcfdd9a096bc8e84dc44a69c97cff3f"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.493293 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" event={"ID":"f46571c0-32a7-4202-b308-54446adfe3d1","Type":"ContainerStarted","Data":"8788d12997f267600736e4fce00b44359c635d7d420fa770f3f6d7432b3b8f50"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.496631 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" event={"ID":"f7b65ead-d9d5-4f06-80de-ae5c85dcbd45","Type":"ContainerStarted","Data":"51e84390bb0f79dfbee9dfddda71ec70dbfd58604c3507c7d3192e847b477109"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.496657 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" event={"ID":"f7b65ead-d9d5-4f06-80de-ae5c85dcbd45","Type":"ContainerStarted","Data":"430aa08a7a33247f14ceb7618d02258085eb8c9333541a4445f6d48079aa83f1"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.497546 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.506101 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" event={"ID":"e8242a68-738b-4fcd-abaa-22a18790ea28","Type":"ContainerStarted","Data":"44c7ec158f623b80ebd3eedb55a3186add8485d61d7206dc75a80e15081e8c91"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.506136 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" event={"ID":"e8242a68-738b-4fcd-abaa-22a18790ea28","Type":"ContainerStarted","Data":"0230e193ec39836bbe1b4e5609ac6338a1be0d69138f072ff68b6cbec32871e4"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.506516 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.524785 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" podStartSLOduration=3.4714086330000002 podStartE2EDuration="16.524759821s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.984211545 +0000 UTC m=+994.890959057" lastFinishedPulling="2025-10-08 14:17:48.037562753 +0000 UTC m=+1007.944310245" observedRunningTime="2025-10-08 14:17:49.523381823 +0000 UTC m=+1009.430129315" watchObservedRunningTime="2025-10-08 14:17:49.524759821 +0000 UTC m=+1009.431507313" Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.535180 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" event={"ID":"0aa40208-3985-4820-809f-ca1f94b05036","Type":"ContainerStarted","Data":"79aba0f9691a640b08509f005f1e37e9252fc67231c588ce4ec304129b703036"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.542351 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5" event={"ID":"8bc03843-9303-451e-b933-c450aaad7401","Type":"ContainerStarted","Data":"917b3426e9a3e6662ff699f9c61d59135aa808f655dd4325573eea77dff93daa"} Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.552784 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" podStartSLOduration=3.803151482 podStartE2EDuration="16.55276244s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.55783668 +0000 UTC m=+995.464584172" lastFinishedPulling="2025-10-08 14:17:48.307447638 +0000 UTC m=+1008.214195130" observedRunningTime="2025-10-08 14:17:49.543310183 +0000 UTC m=+1009.450057675" watchObservedRunningTime="2025-10-08 14:17:49.55276244 +0000 UTC m=+1009.459509932" Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.565355 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5" podStartSLOduration=3.9338258440000002 podStartE2EDuration="16.5653385s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.40650517 +0000 UTC m=+995.313252662" lastFinishedPulling="2025-10-08 14:17:48.038017836 +0000 UTC m=+1007.944765318" observedRunningTime="2025-10-08 14:17:49.563380777 +0000 UTC m=+1009.470128269" watchObservedRunningTime="2025-10-08 14:17:49.5653385 +0000 UTC m=+1009.472085992" Oct 08 14:17:49 crc kubenswrapper[4789]: I1008 14:17:49.565920 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" event={"ID":"78f0cb36-5bad-4021-b464-c5da9318404a","Type":"ContainerStarted","Data":"5614518cdb70336b0ab3b3c1f0c266c9cdec67e9ca77258a66016db8f9901e0d"} Oct 08 14:17:50 crc kubenswrapper[4789]: I1008 14:17:50.574443 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" event={"ID":"78f0cb36-5bad-4021-b464-c5da9318404a","Type":"ContainerStarted","Data":"51ab8c41d80ecd71c7942800d0d02c691bae7b13c5ff02791c3e206f60c22580"} Oct 08 14:17:50 crc kubenswrapper[4789]: I1008 14:17:50.595633 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" podStartSLOduration=5.250167988 podStartE2EDuration="18.595609822s" podCreationTimestamp="2025-10-08 14:17:32 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.552574697 +0000 UTC m=+994.459322189" lastFinishedPulling="2025-10-08 14:17:47.898016531 +0000 UTC m=+1007.804764023" observedRunningTime="2025-10-08 14:17:50.591167992 +0000 UTC m=+1010.497915484" watchObservedRunningTime="2025-10-08 14:17:50.595609822 +0000 UTC m=+1010.502357344" Oct 08 14:17:51 crc kubenswrapper[4789]: I1008 14:17:51.582707 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" event={"ID":"0aa40208-3985-4820-809f-ca1f94b05036","Type":"ContainerStarted","Data":"0d77d7a86289e55c228c403042a5824e8de9f80c5bb5114c9e054b0c80a6966a"} Oct 08 14:17:51 crc kubenswrapper[4789]: I1008 14:17:51.583114 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" Oct 08 14:17:53 crc kubenswrapper[4789]: I1008 14:17:53.313382 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-k8txt" Oct 08 14:17:53 crc kubenswrapper[4789]: I1008 14:17:53.598669 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" event={"ID":"c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4","Type":"ContainerStarted","Data":"b29d48c53398bcee4821fbf8cb93b1ec5c63990a61c17e9c81bc9ffbf4c9b5ad"} Oct 08 14:17:53 crc kubenswrapper[4789]: I1008 14:17:53.871255 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-bvjfl" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.024318 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-wl7d2" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.614768 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" event={"ID":"3c0f54fe-d45f-44c2-9f09-4065c9621903","Type":"ContainerStarted","Data":"667130a7bd462b6454151dbe1acf5efeae6f40f1186c2e9eb124dc5bd3bb1153"} Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.615225 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.617336 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" event={"ID":"c9e3e3c8-447f-4880-b54b-1d5601271216","Type":"ContainerStarted","Data":"a099b22e9e56e2bdc87cae4cc47c8a2b7eb0e75e0f760d1485a1bf38ca81945d"} Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.617536 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.617716 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.620544 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" event={"ID":"bff33916-174d-4fd4-a5e4-91f7cacb6e52","Type":"ContainerStarted","Data":"14f5a5fdb97475d4db344959a96e66da7d7170bfec2c4a8e5709e55150cfd528"} Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.620827 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.620975 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.622438 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.622839 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" event={"ID":"5578f70a-ad1b-47f8-a6ce-5dc726788ded","Type":"ContainerStarted","Data":"9f6662b14482bccbd1892fecb588b7ef9aff945bd2a57e419329fe61e1872323"} Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.623335 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.624529 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.628777 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" event={"ID":"f46571c0-32a7-4202-b308-54446adfe3d1","Type":"ContainerStarted","Data":"9f1dd6433022f69c113d476e8fd70edb4361615fe556c13b3b4ce3c9917f69e8"} Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.629032 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.631043 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.631292 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-7gsbz" podStartSLOduration=9.994331076 podStartE2EDuration="22.63128103s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.400658901 +0000 UTC m=+995.307406393" lastFinishedPulling="2025-10-08 14:17:48.037608855 +0000 UTC m=+1007.944356347" observedRunningTime="2025-10-08 14:17:55.629978555 +0000 UTC m=+1015.536726067" watchObservedRunningTime="2025-10-08 14:17:55.63128103 +0000 UTC m=+1015.538028522" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.634091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" event={"ID":"08eb1945-dc5b-4f43-9fad-cd2b7e3102be","Type":"ContainerStarted","Data":"928c6244eecbc752a93e4a59b3a496e65a0174fadb25303a3f7cb900744cdff4"} Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.634431 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.635828 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.636422 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" event={"ID":"0d8d4170-6f10-4d04-b443-297aada2dc10","Type":"ContainerStarted","Data":"16a03667bbef5b1e279cbe80a75458a97f9aacc572fa33013004a07a9e25bc75"} Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.637015 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.638573 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.644923 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" event={"ID":"8e51126b-8f41-4627-9f68-fb1255bda5bd","Type":"ContainerStarted","Data":"12788c4c667b452b435c86afab4fdc7386d9327793ffdf0861eea343005b7d5d"} Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.645226 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.647722 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.648341 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" event={"ID":"20bc6d13-f85d-423a-9782-097696fe3a0a","Type":"ContainerStarted","Data":"3d7abd7ed6e8af5625839de6226128d628bc52b8e76287e94805448efd15f0f0"} Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.648973 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.649207 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-6f9x9" podStartSLOduration=9.324806341 podStartE2EDuration="22.649188535s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.985428958 +0000 UTC m=+994.892176450" lastFinishedPulling="2025-10-08 14:17:48.309811152 +0000 UTC m=+1008.216558644" observedRunningTime="2025-10-08 14:17:55.644395325 +0000 UTC m=+1015.551142827" watchObservedRunningTime="2025-10-08 14:17:55.649188535 +0000 UTC m=+1015.555936027" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.650308 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.712156 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-w5bgc" podStartSLOduration=14.237327869 podStartE2EDuration="22.712137191s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.577967375 +0000 UTC m=+994.484714857" lastFinishedPulling="2025-10-08 14:17:43.052776687 +0000 UTC m=+1002.959524179" observedRunningTime="2025-10-08 14:17:55.707168736 +0000 UTC m=+1015.613916228" watchObservedRunningTime="2025-10-08 14:17:55.712137191 +0000 UTC m=+1015.618884683" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.746495 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-zslsw" podStartSLOduration=9.243632942 podStartE2EDuration="22.746477812s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.81281873 +0000 UTC m=+994.719566222" lastFinishedPulling="2025-10-08 14:17:48.3156636 +0000 UTC m=+1008.222411092" observedRunningTime="2025-10-08 14:17:55.737504649 +0000 UTC m=+1015.644252141" watchObservedRunningTime="2025-10-08 14:17:55.746477812 +0000 UTC m=+1015.653225304" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.772380 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-7trlm" podStartSLOduration=9.365712809 podStartE2EDuration="22.772361493s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.956808412 +0000 UTC m=+994.863555904" lastFinishedPulling="2025-10-08 14:17:48.363457096 +0000 UTC m=+1008.270204588" observedRunningTime="2025-10-08 14:17:55.760316337 +0000 UTC m=+1015.667063829" watchObservedRunningTime="2025-10-08 14:17:55.772361493 +0000 UTC m=+1015.679108985" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.829722 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-74665f6cdc-x52cc" podStartSLOduration=9.925919583 podStartE2EDuration="22.829697957s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.406455799 +0000 UTC m=+995.313203291" lastFinishedPulling="2025-10-08 14:17:48.310234173 +0000 UTC m=+1008.216981665" observedRunningTime="2025-10-08 14:17:55.809185331 +0000 UTC m=+1015.715932823" watchObservedRunningTime="2025-10-08 14:17:55.829697957 +0000 UTC m=+1015.736445449" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.850907 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" podStartSLOduration=10.793662898000001 podStartE2EDuration="23.850891922s" podCreationTimestamp="2025-10-08 14:17:32 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.997577147 +0000 UTC m=+994.904324639" lastFinishedPulling="2025-10-08 14:17:48.054806171 +0000 UTC m=+1007.961553663" observedRunningTime="2025-10-08 14:17:55.82796981 +0000 UTC m=+1015.734717302" watchObservedRunningTime="2025-10-08 14:17:55.850891922 +0000 UTC m=+1015.757639414" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.904792 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-pzg7c" podStartSLOduration=9.589077233 podStartE2EDuration="22.904776352s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.994550475 +0000 UTC m=+994.901297967" lastFinishedPulling="2025-10-08 14:17:48.310249584 +0000 UTC m=+1008.216997086" observedRunningTime="2025-10-08 14:17:55.866858064 +0000 UTC m=+1015.773605556" watchObservedRunningTime="2025-10-08 14:17:55.904776352 +0000 UTC m=+1015.811523844" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.908626 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-77hn5" podStartSLOduration=10.410123823 podStartE2EDuration="23.908613906s" podCreationTimestamp="2025-10-08 14:17:32 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.811389691 +0000 UTC m=+994.718137183" lastFinishedPulling="2025-10-08 14:17:48.309879774 +0000 UTC m=+1008.216627266" observedRunningTime="2025-10-08 14:17:55.904376391 +0000 UTC m=+1015.811123883" watchObservedRunningTime="2025-10-08 14:17:55.908613906 +0000 UTC m=+1015.815361398" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.933108 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" podStartSLOduration=9.578336062 podStartE2EDuration="22.933088129s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.955641431 +0000 UTC m=+994.862388933" lastFinishedPulling="2025-10-08 14:17:48.310393508 +0000 UTC m=+1008.217141000" observedRunningTime="2025-10-08 14:17:55.932045961 +0000 UTC m=+1015.838793473" watchObservedRunningTime="2025-10-08 14:17:55.933088129 +0000 UTC m=+1015.839835621" Oct 08 14:17:55 crc kubenswrapper[4789]: I1008 14:17:55.966765 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-vbtkd" podStartSLOduration=10.283750839 podStartE2EDuration="23.966745972s" podCreationTimestamp="2025-10-08 14:17:32 +0000 UTC" firstStartedPulling="2025-10-08 14:17:34.6312772 +0000 UTC m=+994.538024692" lastFinishedPulling="2025-10-08 14:17:48.314272333 +0000 UTC m=+1008.221019825" observedRunningTime="2025-10-08 14:17:55.966265369 +0000 UTC m=+1015.873012861" watchObservedRunningTime="2025-10-08 14:17:55.966745972 +0000 UTC m=+1015.873493464" Oct 08 14:17:57 crc kubenswrapper[4789]: I1008 14:17:57.672754 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" event={"ID":"00c51d73-7d74-484a-8773-41cbb40c69f9","Type":"ContainerStarted","Data":"bb84509bb0a8c58c82ca60be187a009277ffa85104f3b18a689b2cbb859dc5c1"} Oct 08 14:17:57 crc kubenswrapper[4789]: I1008 14:17:57.677793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" event={"ID":"67d8c63c-cc55-4be0-bea4-8371ac8e333b","Type":"ContainerStarted","Data":"90d9e3ae9dcb21116d098193aa8395e601b28f0e245f66346fabf10327041983"} Oct 08 14:17:57 crc kubenswrapper[4789]: I1008 14:17:57.678530 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" Oct 08 14:17:57 crc kubenswrapper[4789]: I1008 14:17:57.680164 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" event={"ID":"4e7cd567-305f-417d-bac8-7bf4caf22693","Type":"ContainerStarted","Data":"934f9796d364c5737b0099c5ddbba9deab03430a13a14f1f7f70fa075e26325a"} Oct 08 14:17:57 crc kubenswrapper[4789]: I1008 14:17:57.680503 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" Oct 08 14:17:57 crc kubenswrapper[4789]: I1008 14:17:57.694693 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" podStartSLOduration=4.250428031 podStartE2EDuration="25.694675571s" podCreationTimestamp="2025-10-08 14:17:32 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.46482213 +0000 UTC m=+995.371569622" lastFinishedPulling="2025-10-08 14:17:56.90906967 +0000 UTC m=+1016.815817162" observedRunningTime="2025-10-08 14:17:57.68948809 +0000 UTC m=+1017.596235582" watchObservedRunningTime="2025-10-08 14:17:57.694675571 +0000 UTC m=+1017.601423063" Oct 08 14:17:57 crc kubenswrapper[4789]: I1008 14:17:57.709515 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" podStartSLOduration=3.320784177 podStartE2EDuration="24.709495343s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.520037526 +0000 UTC m=+995.426785018" lastFinishedPulling="2025-10-08 14:17:56.908748692 +0000 UTC m=+1016.815496184" observedRunningTime="2025-10-08 14:17:57.704812196 +0000 UTC m=+1017.611559708" watchObservedRunningTime="2025-10-08 14:17:57.709495343 +0000 UTC m=+1017.616242835" Oct 08 14:17:57 crc kubenswrapper[4789]: I1008 14:17:57.720762 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" podStartSLOduration=3.244070978 podStartE2EDuration="24.720746067s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.465390276 +0000 UTC m=+995.372137768" lastFinishedPulling="2025-10-08 14:17:56.942065365 +0000 UTC m=+1016.848812857" observedRunningTime="2025-10-08 14:17:57.718125586 +0000 UTC m=+1017.624873088" watchObservedRunningTime="2025-10-08 14:17:57.720746067 +0000 UTC m=+1017.627493559" Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.695812 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" event={"ID":"6f893fd6-e5cc-42b9-ac46-77567818c0ef","Type":"ContainerStarted","Data":"f3e6332c8092d036a28012114c5a5acbb5020ad62bd1c24d042252ae828d6d13"} Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.696672 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.697637 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" event={"ID":"8bdbf6d1-353c-42ba-80df-e60d6b424f55","Type":"ContainerStarted","Data":"e0f22817daba853566d933b7a130f9dc804e84a211757d72f2f8dda5b253e020"} Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.697900 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.699888 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" event={"ID":"bc485c37-780a-4509-b3f3-aaf1714a9edb","Type":"ContainerStarted","Data":"12434efb56633a543273834e1f5af402ab36c50c92051a0ce80bcecda3bac197"} Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.700116 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.702175 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" event={"ID":"8efa82a4-efc9-4abe-b373-8828def5bb61","Type":"ContainerStarted","Data":"50a5a38357ac0a4d3362e5f571a074a07ca2d7a46146b305e0ba11eb12d1630e"} Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.702381 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.726336 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" podStartSLOduration=3.336632688 podStartE2EDuration="26.726320033s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.520377745 +0000 UTC m=+995.427125237" lastFinishedPulling="2025-10-08 14:17:58.91006509 +0000 UTC m=+1018.816812582" observedRunningTime="2025-10-08 14:17:59.720733361 +0000 UTC m=+1019.627480853" watchObservedRunningTime="2025-10-08 14:17:59.726320033 +0000 UTC m=+1019.633067525" Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.737475 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" podStartSLOduration=3.246520636 podStartE2EDuration="26.737458675s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.475032527 +0000 UTC m=+995.381780019" lastFinishedPulling="2025-10-08 14:17:58.965970566 +0000 UTC m=+1018.872718058" observedRunningTime="2025-10-08 14:17:59.736061137 +0000 UTC m=+1019.642808649" watchObservedRunningTime="2025-10-08 14:17:59.737458675 +0000 UTC m=+1019.644206167" Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.754907 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" podStartSLOduration=3.3087137220000002 podStartE2EDuration="26.754891437s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.44154166 +0000 UTC m=+995.348289152" lastFinishedPulling="2025-10-08 14:17:58.887719375 +0000 UTC m=+1018.794466867" observedRunningTime="2025-10-08 14:17:59.749499821 +0000 UTC m=+1019.656247313" watchObservedRunningTime="2025-10-08 14:17:59.754891437 +0000 UTC m=+1019.661638929" Oct 08 14:17:59 crc kubenswrapper[4789]: I1008 14:17:59.768920 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" podStartSLOduration=3.353489105 podStartE2EDuration="26.768903457s" podCreationTimestamp="2025-10-08 14:17:33 +0000 UTC" firstStartedPulling="2025-10-08 14:17:35.472628422 +0000 UTC m=+995.379375914" lastFinishedPulling="2025-10-08 14:17:58.888042764 +0000 UTC m=+1018.794790266" observedRunningTime="2025-10-08 14:17:59.765478634 +0000 UTC m=+1019.672226136" watchObservedRunningTime="2025-10-08 14:17:59.768903457 +0000 UTC m=+1019.675650959" Oct 08 14:18:03 crc kubenswrapper[4789]: I1008 14:18:03.637316 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" Oct 08 14:18:03 crc kubenswrapper[4789]: I1008 14:18:03.639430 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ptjbw" Oct 08 14:18:03 crc kubenswrapper[4789]: I1008 14:18:03.655197 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" Oct 08 14:18:03 crc kubenswrapper[4789]: I1008 14:18:03.658840 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-tggcc" Oct 08 14:18:03 crc kubenswrapper[4789]: I1008 14:18:03.821408 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" Oct 08 14:18:03 crc kubenswrapper[4789]: I1008 14:18:03.823159 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-2j5bd" Oct 08 14:18:03 crc kubenswrapper[4789]: I1008 14:18:03.959640 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-2kzff" Oct 08 14:18:04 crc kubenswrapper[4789]: I1008 14:18:04.003038 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-775776c574-mgjfp" Oct 08 14:18:04 crc kubenswrapper[4789]: I1008 14:18:04.064781 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4xgrb" Oct 08 14:18:04 crc kubenswrapper[4789]: I1008 14:18:04.282118 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-nfllk" Oct 08 14:18:04 crc kubenswrapper[4789]: I1008 14:18:04.314372 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c7dbfbcbd-7xrj5" Oct 08 14:18:04 crc kubenswrapper[4789]: I1008 14:18:04.587539 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.224580 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67f9b9c8c7-8glrx"] Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.226416 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.228308 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.228390 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.228500 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.228587 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-nxmjz" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.235973 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67f9b9c8c7-8glrx"] Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.283975 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8fbc6f7c-m2f76"] Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.285527 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.290736 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.294502 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8fbc6f7c-m2f76"] Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.350512 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx4rp\" (UniqueName: \"kubernetes.io/projected/55261335-33b4-419d-92bc-e5442e647e5d-kube-api-access-nx4rp\") pod \"dnsmasq-dns-67f9b9c8c7-8glrx\" (UID: \"55261335-33b4-419d-92bc-e5442e647e5d\") " pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.350559 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55261335-33b4-419d-92bc-e5442e647e5d-config\") pod \"dnsmasq-dns-67f9b9c8c7-8glrx\" (UID: \"55261335-33b4-419d-92bc-e5442e647e5d\") " pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.451955 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxgsf\" (UniqueName: \"kubernetes.io/projected/12b4276a-e078-439c-a236-0069eadc0a78-kube-api-access-mxgsf\") pod \"dnsmasq-dns-8fbc6f7c-m2f76\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.452024 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-config\") pod \"dnsmasq-dns-8fbc6f7c-m2f76\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.452044 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-dns-svc\") pod \"dnsmasq-dns-8fbc6f7c-m2f76\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.452082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx4rp\" (UniqueName: \"kubernetes.io/projected/55261335-33b4-419d-92bc-e5442e647e5d-kube-api-access-nx4rp\") pod \"dnsmasq-dns-67f9b9c8c7-8glrx\" (UID: \"55261335-33b4-419d-92bc-e5442e647e5d\") " pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.452105 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55261335-33b4-419d-92bc-e5442e647e5d-config\") pod \"dnsmasq-dns-67f9b9c8c7-8glrx\" (UID: \"55261335-33b4-419d-92bc-e5442e647e5d\") " pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.453344 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55261335-33b4-419d-92bc-e5442e647e5d-config\") pod \"dnsmasq-dns-67f9b9c8c7-8glrx\" (UID: \"55261335-33b4-419d-92bc-e5442e647e5d\") " pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.476674 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx4rp\" (UniqueName: \"kubernetes.io/projected/55261335-33b4-419d-92bc-e5442e647e5d-kube-api-access-nx4rp\") pod \"dnsmasq-dns-67f9b9c8c7-8glrx\" (UID: \"55261335-33b4-419d-92bc-e5442e647e5d\") " pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.547123 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.553663 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxgsf\" (UniqueName: \"kubernetes.io/projected/12b4276a-e078-439c-a236-0069eadc0a78-kube-api-access-mxgsf\") pod \"dnsmasq-dns-8fbc6f7c-m2f76\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.553949 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-config\") pod \"dnsmasq-dns-8fbc6f7c-m2f76\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.554052 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-dns-svc\") pod \"dnsmasq-dns-8fbc6f7c-m2f76\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.554807 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-config\") pod \"dnsmasq-dns-8fbc6f7c-m2f76\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.556081 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-dns-svc\") pod \"dnsmasq-dns-8fbc6f7c-m2f76\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.572146 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxgsf\" (UniqueName: \"kubernetes.io/projected/12b4276a-e078-439c-a236-0069eadc0a78-kube-api-access-mxgsf\") pod \"dnsmasq-dns-8fbc6f7c-m2f76\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:23 crc kubenswrapper[4789]: I1008 14:18:23.606073 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:24 crc kubenswrapper[4789]: I1008 14:18:24.009433 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67f9b9c8c7-8glrx"] Oct 08 14:18:24 crc kubenswrapper[4789]: I1008 14:18:24.096583 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8fbc6f7c-m2f76"] Oct 08 14:18:24 crc kubenswrapper[4789]: W1008 14:18:24.100559 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12b4276a_e078_439c_a236_0069eadc0a78.slice/crio-e5f655075ab8c52d25f16eb0e865efb8e4fe27a9a3628579f87f7c2bdfe12378 WatchSource:0}: Error finding container e5f655075ab8c52d25f16eb0e865efb8e4fe27a9a3628579f87f7c2bdfe12378: Status 404 returned error can't find the container with id e5f655075ab8c52d25f16eb0e865efb8e4fe27a9a3628579f87f7c2bdfe12378 Oct 08 14:18:24 crc kubenswrapper[4789]: I1008 14:18:24.941500 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" event={"ID":"55261335-33b4-419d-92bc-e5442e647e5d","Type":"ContainerStarted","Data":"bbb943a5b1eb5af4711a705ac21de6580a068e79e0cbae1e0ee581a748ecfca7"} Oct 08 14:18:24 crc kubenswrapper[4789]: I1008 14:18:24.943559 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" event={"ID":"12b4276a-e078-439c-a236-0069eadc0a78","Type":"ContainerStarted","Data":"e5f655075ab8c52d25f16eb0e865efb8e4fe27a9a3628579f87f7c2bdfe12378"} Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.253589 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67f9b9c8c7-8glrx"] Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.276064 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68988576c-9hdxg"] Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.277222 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.289192 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68988576c-9hdxg"] Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.420425 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsp6k\" (UniqueName: \"kubernetes.io/projected/28abfdd3-7f5a-4864-a827-d6c1693ee243-kube-api-access-gsp6k\") pod \"dnsmasq-dns-68988576c-9hdxg\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.420477 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-config\") pod \"dnsmasq-dns-68988576c-9hdxg\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.420563 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-dns-svc\") pod \"dnsmasq-dns-68988576c-9hdxg\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.521615 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsp6k\" (UniqueName: \"kubernetes.io/projected/28abfdd3-7f5a-4864-a827-d6c1693ee243-kube-api-access-gsp6k\") pod \"dnsmasq-dns-68988576c-9hdxg\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.521683 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-config\") pod \"dnsmasq-dns-68988576c-9hdxg\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.521746 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-dns-svc\") pod \"dnsmasq-dns-68988576c-9hdxg\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.522760 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-dns-svc\") pod \"dnsmasq-dns-68988576c-9hdxg\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.523751 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-config\") pod \"dnsmasq-dns-68988576c-9hdxg\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.550853 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsp6k\" (UniqueName: \"kubernetes.io/projected/28abfdd3-7f5a-4864-a827-d6c1693ee243-kube-api-access-gsp6k\") pod \"dnsmasq-dns-68988576c-9hdxg\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.608223 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.651303 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fbc6f7c-m2f76"] Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.728910 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d975bffc9-hjkl7"] Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.731138 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.742665 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d975bffc9-hjkl7"] Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.831562 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-dns-svc\") pod \"dnsmasq-dns-5d975bffc9-hjkl7\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.831918 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-config\") pod \"dnsmasq-dns-5d975bffc9-hjkl7\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.832181 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd6k9\" (UniqueName: \"kubernetes.io/projected/40614905-004f-4a32-89fb-04bc38f13165-kube-api-access-nd6k9\") pod \"dnsmasq-dns-5d975bffc9-hjkl7\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.933177 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-dns-svc\") pod \"dnsmasq-dns-5d975bffc9-hjkl7\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.933224 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-config\") pod \"dnsmasq-dns-5d975bffc9-hjkl7\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.933372 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd6k9\" (UniqueName: \"kubernetes.io/projected/40614905-004f-4a32-89fb-04bc38f13165-kube-api-access-nd6k9\") pod \"dnsmasq-dns-5d975bffc9-hjkl7\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.934536 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-dns-svc\") pod \"dnsmasq-dns-5d975bffc9-hjkl7\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.934589 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-config\") pod \"dnsmasq-dns-5d975bffc9-hjkl7\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:27 crc kubenswrapper[4789]: I1008 14:18:27.982899 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd6k9\" (UniqueName: \"kubernetes.io/projected/40614905-004f-4a32-89fb-04bc38f13165-kube-api-access-nd6k9\") pod \"dnsmasq-dns-5d975bffc9-hjkl7\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.009154 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d975bffc9-hjkl7"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.009660 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.019534 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7547999-h5qmx"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.020833 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.024258 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7547999-h5qmx"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.047100 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-dns-svc\") pod \"dnsmasq-dns-5d7547999-h5qmx\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.047154 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dknt\" (UniqueName: \"kubernetes.io/projected/40054025-fd50-4abc-a3c9-2241ef2be0e9-kube-api-access-7dknt\") pod \"dnsmasq-dns-5d7547999-h5qmx\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.047212 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-config\") pod \"dnsmasq-dns-5d7547999-h5qmx\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.148737 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-dns-svc\") pod \"dnsmasq-dns-5d7547999-h5qmx\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.149097 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dknt\" (UniqueName: \"kubernetes.io/projected/40054025-fd50-4abc-a3c9-2241ef2be0e9-kube-api-access-7dknt\") pod \"dnsmasq-dns-5d7547999-h5qmx\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.149123 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-config\") pod \"dnsmasq-dns-5d7547999-h5qmx\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.150051 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-config\") pod \"dnsmasq-dns-5d7547999-h5qmx\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.150592 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-dns-svc\") pod \"dnsmasq-dns-5d7547999-h5qmx\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.173073 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dknt\" (UniqueName: \"kubernetes.io/projected/40054025-fd50-4abc-a3c9-2241ef2be0e9-kube-api-access-7dknt\") pod \"dnsmasq-dns-5d7547999-h5qmx\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.199354 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68988576c-9hdxg"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.381839 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.466569 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.469556 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.474426 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.474504 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.474612 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.474680 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.474737 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.474818 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.474935 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-xrf5t" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.478384 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.635288 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d975bffc9-hjkl7"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.645633 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7547999-h5qmx"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659648 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659720 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659754 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c9336a8d-1117-4da6-a21b-4070f722adc9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659772 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659792 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659813 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659842 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c9336a8d-1117-4da6-a21b-4070f722adc9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659856 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659872 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7d6z\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-kube-api-access-k7d6z\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659898 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.659917 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761374 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761462 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761524 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761546 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761576 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c9336a8d-1117-4da6-a21b-4070f722adc9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761594 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761613 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761637 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761669 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c9336a8d-1117-4da6-a21b-4070f722adc9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761688 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.761709 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7d6z\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-kube-api-access-k7d6z\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.763160 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.763279 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.763690 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.763952 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.764027 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.766365 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.772163 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c9336a8d-1117-4da6-a21b-4070f722adc9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.772694 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.772804 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.775355 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c9336a8d-1117-4da6-a21b-4070f722adc9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.780576 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7d6z\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-kube-api-access-k7d6z\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.785521 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.797079 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.809531 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.811074 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.813048 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-erlang-cookie" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.813865 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-plugins-conf" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.813908 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-server-dockercfg-c24qq" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.814082 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-server-conf" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.814127 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-config-data" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.814306 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-notifications-svc" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.816455 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-default-user" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.821775 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.964505 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.964573 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55fc0a2c-48c9-4508-ae9d-5350f3298621-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.964605 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.964623 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.964672 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/55fc0a2c-48c9-4508-ae9d-5350f3298621-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.964727 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.964747 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.964766 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/55fc0a2c-48c9-4508-ae9d-5350f3298621-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.964819 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/55fc0a2c-48c9-4508-ae9d-5350f3298621-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.965397 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/55fc0a2c-48c9-4508-ae9d-5350f3298621-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.965467 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkpdp\" (UniqueName: \"kubernetes.io/projected/55fc0a2c-48c9-4508-ae9d-5350f3298621-kube-api-access-lkpdp\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:28 crc kubenswrapper[4789]: I1008 14:18:28.999109 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" event={"ID":"40054025-fd50-4abc-a3c9-2241ef2be0e9","Type":"ContainerStarted","Data":"4deea8a8c2275f9660fbb7bf832c644debfc06335f6bc6b10e672722761265fe"} Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.000525 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68988576c-9hdxg" event={"ID":"28abfdd3-7f5a-4864-a827-d6c1693ee243","Type":"ContainerStarted","Data":"3f2e799a9f7f443ab6c5744c38935b8fbc87b75053280e8be4ae41da87db0cd1"} Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.002429 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" event={"ID":"40614905-004f-4a32-89fb-04bc38f13165","Type":"ContainerStarted","Data":"7e1ad871f96cf7bb9b29938a38b3a8bde150261e042c2977dc1cf23dbbfc2911"} Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067097 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkpdp\" (UniqueName: \"kubernetes.io/projected/55fc0a2c-48c9-4508-ae9d-5350f3298621-kube-api-access-lkpdp\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067155 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067202 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55fc0a2c-48c9-4508-ae9d-5350f3298621-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067227 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067242 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067289 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/55fc0a2c-48c9-4508-ae9d-5350f3298621-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067307 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067322 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067338 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/55fc0a2c-48c9-4508-ae9d-5350f3298621-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067363 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/55fc0a2c-48c9-4508-ae9d-5350f3298621-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.067381 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/55fc0a2c-48c9-4508-ae9d-5350f3298621-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.068276 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.068369 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.069047 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55fc0a2c-48c9-4508-ae9d-5350f3298621-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.069597 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/55fc0a2c-48c9-4508-ae9d-5350f3298621-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.069904 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/55fc0a2c-48c9-4508-ae9d-5350f3298621-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.070835 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.071450 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/55fc0a2c-48c9-4508-ae9d-5350f3298621-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.073139 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/55fc0a2c-48c9-4508-ae9d-5350f3298621-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.073949 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.074524 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/55fc0a2c-48c9-4508-ae9d-5350f3298621-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.085639 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkpdp\" (UniqueName: \"kubernetes.io/projected/55fc0a2c-48c9-4508-ae9d-5350f3298621-kube-api-access-lkpdp\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.095244 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"55fc0a2c-48c9-4508-ae9d-5350f3298621\") " pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.149012 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.149454 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.150388 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.153115 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.153696 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.153842 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.153901 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.154100 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.154218 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.154237 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rmgmp" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.168494 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.249957 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.270785 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.270844 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.270876 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.270899 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-server-conf\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.270943 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-config-data\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.270978 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05192759-43d6-4ef0-b6f0-6284e2435317-pod-info\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.271016 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.271060 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtzwg\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-kube-api-access-wtzwg\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.271088 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05192759-43d6-4ef0-b6f0-6284e2435317-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.271114 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.271141 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.372951 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373106 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-server-conf\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373242 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-config-data\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373285 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05192759-43d6-4ef0-b6f0-6284e2435317-pod-info\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373307 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373353 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtzwg\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-kube-api-access-wtzwg\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373382 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05192759-43d6-4ef0-b6f0-6284e2435317-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373412 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373442 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373473 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373518 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.373789 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.374818 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-config-data\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.375093 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.375479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.375474 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.375775 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-server-conf\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.379574 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.379729 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.382582 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05192759-43d6-4ef0-b6f0-6284e2435317-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.383622 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05192759-43d6-4ef0-b6f0-6284e2435317-pod-info\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.392046 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtzwg\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-kube-api-access-wtzwg\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.403084 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.523296 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 14:18:29 crc kubenswrapper[4789]: I1008 14:18:29.603011 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Oct 08 14:18:30 crc kubenswrapper[4789]: I1008 14:18:30.017796 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c9336a8d-1117-4da6-a21b-4070f722adc9","Type":"ContainerStarted","Data":"f1974ce70aa450a751575356d02ca0d70f7dd4b0cad016edb8124bcf862c705d"} Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.702898 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.706712 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.714461 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4hjx8" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.715053 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.715703 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.718526 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.724125 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.726301 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.731318 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.735621 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.737182 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.740392 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.740479 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-qmrbb" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.740648 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.740768 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.742019 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.824882 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-kolla-config\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.824954 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825064 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32ceb31c-1751-4a04-a91a-b317f3ba3d78-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825159 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825230 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wtpl\" (UniqueName: \"kubernetes.io/projected/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-kube-api-access-4wtpl\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825256 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ceb31c-1751-4a04-a91a-b317f3ba3d78-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825311 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-secrets\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825328 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825405 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-config-data-default\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825459 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/32ceb31c-1751-4a04-a91a-b317f3ba3d78-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825504 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/32ceb31c-1751-4a04-a91a-b317f3ba3d78-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825557 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825576 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8dg6\" (UniqueName: \"kubernetes.io/projected/32ceb31c-1751-4a04-a91a-b317f3ba3d78-kube-api-access-s8dg6\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.825643 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/32ceb31c-1751-4a04-a91a-b317f3ba3d78-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.841749 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.841903 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.842015 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ceb31c-1751-4a04-a91a-b317f3ba3d78-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.842049 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32ceb31c-1751-4a04-a91a-b317f3ba3d78-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943218 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943265 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wtpl\" (UniqueName: \"kubernetes.io/projected/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-kube-api-access-4wtpl\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943282 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ceb31c-1751-4a04-a91a-b317f3ba3d78-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943316 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-secrets\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943339 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943368 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-config-data-default\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943386 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/32ceb31c-1751-4a04-a91a-b317f3ba3d78-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943401 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/32ceb31c-1751-4a04-a91a-b317f3ba3d78-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943417 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.943953 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.944387 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/32ceb31c-1751-4a04-a91a-b317f3ba3d78-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.945444 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-config-data-default\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.945797 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/32ceb31c-1751-4a04-a91a-b317f3ba3d78-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948043 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8dg6\" (UniqueName: \"kubernetes.io/projected/32ceb31c-1751-4a04-a91a-b317f3ba3d78-kube-api-access-s8dg6\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948231 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/32ceb31c-1751-4a04-a91a-b317f3ba3d78-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948250 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948333 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ceb31c-1751-4a04-a91a-b317f3ba3d78-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948358 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32ceb31c-1751-4a04-a91a-b317f3ba3d78-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948405 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-kolla-config\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948431 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948453 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32ceb31c-1751-4a04-a91a-b317f3ba3d78-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.948808 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.949282 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32ceb31c-1751-4a04-a91a-b317f3ba3d78-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.949655 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32ceb31c-1751-4a04-a91a-b317f3ba3d78-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.950140 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-kolla-config\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.950418 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.951351 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/32ceb31c-1751-4a04-a91a-b317f3ba3d78-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.952512 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-secrets\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.952956 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.953042 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ceb31c-1751-4a04-a91a-b317f3ba3d78-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.953061 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ceb31c-1751-4a04-a91a-b317f3ba3d78-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.953662 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.962012 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wtpl\" (UniqueName: \"kubernetes.io/projected/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-kube-api-access-4wtpl\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.963013 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8dg6\" (UniqueName: \"kubernetes.io/projected/32ceb31c-1751-4a04-a91a-b317f3ba3d78-kube-api-access-s8dg6\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.964718 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505cf346-64b1-4de8-8cf6-fca4ddd34cd6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.970994 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"32ceb31c-1751-4a04-a91a-b317f3ba3d78\") " pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:31 crc kubenswrapper[4789]: I1008 14:18:31.979765 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"505cf346-64b1-4de8-8cf6-fca4ddd34cd6\") " pod="openstack/openstack-galera-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.036669 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.057436 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.207121 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.208474 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.211543 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.211599 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-z922x" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.211749 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.225397 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.356432 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9lqk\" (UniqueName: \"kubernetes.io/projected/78e021d8-ac60-4092-9e49-bbb3fff88aa7-kube-api-access-b9lqk\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.356484 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78e021d8-ac60-4092-9e49-bbb3fff88aa7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.356505 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/78e021d8-ac60-4092-9e49-bbb3fff88aa7-kolla-config\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.356787 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/78e021d8-ac60-4092-9e49-bbb3fff88aa7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.356845 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78e021d8-ac60-4092-9e49-bbb3fff88aa7-config-data\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.459135 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/78e021d8-ac60-4092-9e49-bbb3fff88aa7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.459222 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78e021d8-ac60-4092-9e49-bbb3fff88aa7-config-data\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.459279 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9lqk\" (UniqueName: \"kubernetes.io/projected/78e021d8-ac60-4092-9e49-bbb3fff88aa7-kube-api-access-b9lqk\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.459311 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78e021d8-ac60-4092-9e49-bbb3fff88aa7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.459337 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/78e021d8-ac60-4092-9e49-bbb3fff88aa7-kolla-config\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.460112 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/78e021d8-ac60-4092-9e49-bbb3fff88aa7-kolla-config\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.460293 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78e021d8-ac60-4092-9e49-bbb3fff88aa7-config-data\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.465105 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/78e021d8-ac60-4092-9e49-bbb3fff88aa7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.467659 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78e021d8-ac60-4092-9e49-bbb3fff88aa7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.483193 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9lqk\" (UniqueName: \"kubernetes.io/projected/78e021d8-ac60-4092-9e49-bbb3fff88aa7-kube-api-access-b9lqk\") pod \"memcached-0\" (UID: \"78e021d8-ac60-4092-9e49-bbb3fff88aa7\") " pod="openstack/memcached-0" Oct 08 14:18:32 crc kubenswrapper[4789]: I1008 14:18:32.530025 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 14:18:34 crc kubenswrapper[4789]: I1008 14:18:34.466718 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 14:18:34 crc kubenswrapper[4789]: I1008 14:18:34.468055 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 14:18:34 crc kubenswrapper[4789]: I1008 14:18:34.471213 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-z6vpw" Oct 08 14:18:34 crc kubenswrapper[4789]: I1008 14:18:34.478800 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 14:18:34 crc kubenswrapper[4789]: I1008 14:18:34.605703 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtxkm\" (UniqueName: \"kubernetes.io/projected/a4509515-2cf4-40a1-9873-588155ad596e-kube-api-access-jtxkm\") pod \"kube-state-metrics-0\" (UID: \"a4509515-2cf4-40a1-9873-588155ad596e\") " pod="openstack/kube-state-metrics-0" Oct 08 14:18:34 crc kubenswrapper[4789]: I1008 14:18:34.707933 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtxkm\" (UniqueName: \"kubernetes.io/projected/a4509515-2cf4-40a1-9873-588155ad596e-kube-api-access-jtxkm\") pod \"kube-state-metrics-0\" (UID: \"a4509515-2cf4-40a1-9873-588155ad596e\") " pod="openstack/kube-state-metrics-0" Oct 08 14:18:34 crc kubenswrapper[4789]: I1008 14:18:34.757796 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtxkm\" (UniqueName: \"kubernetes.io/projected/a4509515-2cf4-40a1-9873-588155ad596e-kube-api-access-jtxkm\") pod \"kube-state-metrics-0\" (UID: \"a4509515-2cf4-40a1-9873-588155ad596e\") " pod="openstack/kube-state-metrics-0" Oct 08 14:18:34 crc kubenswrapper[4789]: I1008 14:18:34.788373 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.803269 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.805106 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.808925 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.809101 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.809101 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.809197 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.809451 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-khr4j" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.815322 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.829260 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.926924 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.927020 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.927233 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aaeb67a6-3e80-4d6e-9299-47634a566c72-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.927481 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.927549 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-config\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.927619 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.927921 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aaeb67a6-3e80-4d6e-9299-47634a566c72-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:35 crc kubenswrapper[4789]: I1008 14:18:35.927956 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcm95\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-kube-api-access-rcm95\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.030009 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-config\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.030082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.030159 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aaeb67a6-3e80-4d6e-9299-47634a566c72-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.030182 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcm95\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-kube-api-access-rcm95\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.030247 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.030276 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.030314 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aaeb67a6-3e80-4d6e-9299-47634a566c72-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.030371 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.031008 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aaeb67a6-3e80-4d6e-9299-47634a566c72-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.035818 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.036321 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.039571 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-config\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.039584 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.039817 4789 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.039869 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f75c8fbff2adebf98b3c9ad4427807236950b84a4f856fba0192ffb272b3d6c1/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.040978 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aaeb67a6-3e80-4d6e-9299-47634a566c72-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.049415 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcm95\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-kube-api-access-rcm95\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.070435 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:36 crc kubenswrapper[4789]: I1008 14:18:36.129148 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.803397 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.805506 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.811398 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.811471 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-9cm62" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.811415 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.812561 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.812722 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.813739 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.866769 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zrdhb"] Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.868814 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.870662 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-9lztj" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.871139 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.873231 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.878545 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hhxm\" (UniqueName: \"kubernetes.io/projected/a1a9507e-04fe-4746-a31b-5ad097c48602-kube-api-access-9hhxm\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.878594 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1a9507e-04fe-4746-a31b-5ad097c48602-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.878623 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.878638 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a1a9507e-04fe-4746-a31b-5ad097c48602-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.878660 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a9507e-04fe-4746-a31b-5ad097c48602-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.878724 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a9507e-04fe-4746-a31b-5ad097c48602-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.878823 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a9507e-04fe-4746-a31b-5ad097c48602-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.878886 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a9507e-04fe-4746-a31b-5ad097c48602-config\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.889519 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-tkv8v"] Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.893958 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.912197 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrdhb"] Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.928093 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-tkv8v"] Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.981746 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e2285a3-7296-490c-8649-adeae9a0cabc-scripts\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.981807 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ddb1cc-17c9-4f38-ac8a-d78d10798178-combined-ca-bundle\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.981842 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91ddb1cc-17c9-4f38-ac8a-d78d10798178-var-log-ovn\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.982105 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a9507e-04fe-4746-a31b-5ad097c48602-config\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.982215 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-var-run\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.982388 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnzfz\" (UniqueName: \"kubernetes.io/projected/5e2285a3-7296-490c-8649-adeae9a0cabc-kube-api-access-hnzfz\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.982606 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hhxm\" (UniqueName: \"kubernetes.io/projected/a1a9507e-04fe-4746-a31b-5ad097c48602-kube-api-access-9hhxm\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.982641 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91ddb1cc-17c9-4f38-ac8a-d78d10798178-var-run\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.982676 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1a9507e-04fe-4746-a31b-5ad097c48602-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.982706 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.982729 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a1a9507e-04fe-4746-a31b-5ad097c48602-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.983242 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.983760 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a9507e-04fe-4746-a31b-5ad097c48602-config\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.984252 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a1a9507e-04fe-4746-a31b-5ad097c48602-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.984839 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1a9507e-04fe-4746-a31b-5ad097c48602-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992579 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ddb1cc-17c9-4f38-ac8a-d78d10798178-ovn-controller-tls-certs\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992627 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a9507e-04fe-4746-a31b-5ad097c48602-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992699 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-var-lib\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992728 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-etc-ovs\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992762 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-var-log\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992786 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvx86\" (UniqueName: \"kubernetes.io/projected/91ddb1cc-17c9-4f38-ac8a-d78d10798178-kube-api-access-qvx86\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992838 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91ddb1cc-17c9-4f38-ac8a-d78d10798178-var-run-ovn\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992869 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a9507e-04fe-4746-a31b-5ad097c48602-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992894 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91ddb1cc-17c9-4f38-ac8a-d78d10798178-scripts\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:37 crc kubenswrapper[4789]: I1008 14:18:37.992914 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a9507e-04fe-4746-a31b-5ad097c48602-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.015527 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a9507e-04fe-4746-a31b-5ad097c48602-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.015713 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a9507e-04fe-4746-a31b-5ad097c48602-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.024628 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a9507e-04fe-4746-a31b-5ad097c48602-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.026855 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.027040 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hhxm\" (UniqueName: \"kubernetes.io/projected/a1a9507e-04fe-4746-a31b-5ad097c48602-kube-api-access-9hhxm\") pod \"ovsdbserver-nb-0\" (UID: \"a1a9507e-04fe-4746-a31b-5ad097c48602\") " pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.094825 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnzfz\" (UniqueName: \"kubernetes.io/projected/5e2285a3-7296-490c-8649-adeae9a0cabc-kube-api-access-hnzfz\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.094879 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91ddb1cc-17c9-4f38-ac8a-d78d10798178-var-run\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.094909 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ddb1cc-17c9-4f38-ac8a-d78d10798178-ovn-controller-tls-certs\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.094937 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-var-lib\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.094952 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-etc-ovs\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.094976 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-var-log\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095053 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvx86\" (UniqueName: \"kubernetes.io/projected/91ddb1cc-17c9-4f38-ac8a-d78d10798178-kube-api-access-qvx86\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91ddb1cc-17c9-4f38-ac8a-d78d10798178-var-run-ovn\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095104 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91ddb1cc-17c9-4f38-ac8a-d78d10798178-scripts\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095128 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ddb1cc-17c9-4f38-ac8a-d78d10798178-combined-ca-bundle\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095144 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e2285a3-7296-490c-8649-adeae9a0cabc-scripts\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095165 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91ddb1cc-17c9-4f38-ac8a-d78d10798178-var-log-ovn\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095208 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-var-run\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095725 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91ddb1cc-17c9-4f38-ac8a-d78d10798178-var-run\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095849 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91ddb1cc-17c9-4f38-ac8a-d78d10798178-var-run-ovn\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.096035 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-var-lib\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.096100 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-var-run\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.096130 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-etc-ovs\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.095849 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/5e2285a3-7296-490c-8649-adeae9a0cabc-var-log\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.096163 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91ddb1cc-17c9-4f38-ac8a-d78d10798178-var-log-ovn\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.098098 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e2285a3-7296-490c-8649-adeae9a0cabc-scripts\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.098171 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"55fc0a2c-48c9-4508-ae9d-5350f3298621","Type":"ContainerStarted","Data":"b3f8b86a5b7ed7eb61073da75356e2750352e6eb93c4a2f104fcd4ba1bef2e7f"} Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.099344 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91ddb1cc-17c9-4f38-ac8a-d78d10798178-scripts\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.099455 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ddb1cc-17c9-4f38-ac8a-d78d10798178-combined-ca-bundle\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.103742 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ddb1cc-17c9-4f38-ac8a-d78d10798178-ovn-controller-tls-certs\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.118084 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvx86\" (UniqueName: \"kubernetes.io/projected/91ddb1cc-17c9-4f38-ac8a-d78d10798178-kube-api-access-qvx86\") pod \"ovn-controller-zrdhb\" (UID: \"91ddb1cc-17c9-4f38-ac8a-d78d10798178\") " pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.118429 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnzfz\" (UniqueName: \"kubernetes.io/projected/5e2285a3-7296-490c-8649-adeae9a0cabc-kube-api-access-hnzfz\") pod \"ovn-controller-ovs-tkv8v\" (UID: \"5e2285a3-7296-490c-8649-adeae9a0cabc\") " pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.124641 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.202826 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:38 crc kubenswrapper[4789]: I1008 14:18:38.219060 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.815505 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.819166 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.821186 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.822310 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-h8phv" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.823663 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.824246 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.825873 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.864701 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh9m7\" (UniqueName: \"kubernetes.io/projected/70dc7aa5-b81d-4150-b1dd-57ec365e495c-kube-api-access-xh9m7\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.864793 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/70dc7aa5-b81d-4150-b1dd-57ec365e495c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.864842 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/70dc7aa5-b81d-4150-b1dd-57ec365e495c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.864870 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.864930 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70dc7aa5-b81d-4150-b1dd-57ec365e495c-config\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.864974 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/70dc7aa5-b81d-4150-b1dd-57ec365e495c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.865034 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70dc7aa5-b81d-4150-b1dd-57ec365e495c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.865058 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70dc7aa5-b81d-4150-b1dd-57ec365e495c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.966534 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/70dc7aa5-b81d-4150-b1dd-57ec365e495c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.966586 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.966629 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70dc7aa5-b81d-4150-b1dd-57ec365e495c-config\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.967006 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.967242 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/70dc7aa5-b81d-4150-b1dd-57ec365e495c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.967478 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70dc7aa5-b81d-4150-b1dd-57ec365e495c-config\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.967522 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/70dc7aa5-b81d-4150-b1dd-57ec365e495c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.967574 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70dc7aa5-b81d-4150-b1dd-57ec365e495c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.967597 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70dc7aa5-b81d-4150-b1dd-57ec365e495c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.967999 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh9m7\" (UniqueName: \"kubernetes.io/projected/70dc7aa5-b81d-4150-b1dd-57ec365e495c-kube-api-access-xh9m7\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.968037 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/70dc7aa5-b81d-4150-b1dd-57ec365e495c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.969163 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70dc7aa5-b81d-4150-b1dd-57ec365e495c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.974464 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/70dc7aa5-b81d-4150-b1dd-57ec365e495c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.974574 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70dc7aa5-b81d-4150-b1dd-57ec365e495c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.974972 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/70dc7aa5-b81d-4150-b1dd-57ec365e495c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:41 crc kubenswrapper[4789]: I1008 14:18:41.997011 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh9m7\" (UniqueName: \"kubernetes.io/projected/70dc7aa5-b81d-4150-b1dd-57ec365e495c-kube-api-access-xh9m7\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:42 crc kubenswrapper[4789]: I1008 14:18:42.022245 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"70dc7aa5-b81d-4150-b1dd-57ec365e495c\") " pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:42 crc kubenswrapper[4789]: I1008 14:18:42.148054 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 14:18:45 crc kubenswrapper[4789]: I1008 14:18:45.693033 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 14:18:46 crc kubenswrapper[4789]: I1008 14:18:46.901677 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 14:18:47 crc kubenswrapper[4789]: W1008 14:18:47.156830 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78e021d8_ac60_4092_9e49_bbb3fff88aa7.slice/crio-73796ae8cdec0074fddbf37404297ce8b1a664d9c91b84a8b1395860338e83aa WatchSource:0}: Error finding container 73796ae8cdec0074fddbf37404297ce8b1a664d9c91b84a8b1395860338e83aa: Status 404 returned error can't find the container with id 73796ae8cdec0074fddbf37404297ce8b1a664d9c91b84a8b1395860338e83aa Oct 08 14:18:47 crc kubenswrapper[4789]: I1008 14:18:47.168795 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"78e021d8-ac60-4092-9e49-bbb3fff88aa7","Type":"ContainerStarted","Data":"73796ae8cdec0074fddbf37404297ce8b1a664d9c91b84a8b1395860338e83aa"} Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.209623 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.209675 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.209788 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.222:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mxgsf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-8fbc6f7c-m2f76_openstack(12b4276a-e078-439c-a236-0069eadc0a78): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.211216 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" podUID="12b4276a-e078-439c-a236-0069eadc0a78" Oct 08 14:18:47 crc kubenswrapper[4789]: W1008 14:18:47.222254 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod505cf346_64b1_4de8_8cf6_fca4ddd34cd6.slice/crio-beb36e4f0acca0233969af488745b3d1ad9b8409d137ddee5d9c750d933b95b2 WatchSource:0}: Error finding container beb36e4f0acca0233969af488745b3d1ad9b8409d137ddee5d9c750d933b95b2: Status 404 returned error can't find the container with id beb36e4f0acca0233969af488745b3d1ad9b8409d137ddee5d9c750d933b95b2 Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.289060 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.289449 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.289618 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.222:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gsp6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-68988576c-9hdxg_openstack(28abfdd3-7f5a-4864-a827-d6c1693ee243): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.291201 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-68988576c-9hdxg" podUID="28abfdd3-7f5a-4864-a827-d6c1693ee243" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.325152 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.325206 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.325319 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.222:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nx4rp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-67f9b9c8c7-8glrx_openstack(55261335-33b4-419d-92bc-e5442e647e5d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:18:47 crc kubenswrapper[4789]: E1008 14:18:47.327294 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" podUID="55261335-33b4-419d-92bc-e5442e647e5d" Oct 08 14:18:47 crc kubenswrapper[4789]: I1008 14:18:47.557300 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 14:18:47 crc kubenswrapper[4789]: I1008 14:18:47.751236 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 14:18:47 crc kubenswrapper[4789]: W1008 14:18:47.757453 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32ceb31c_1751_4a04_a91a_b317f3ba3d78.slice/crio-19e2eb4a892a44ae780f4aaced1d0727cb7c7e31b881c78918caedaaa97693ed WatchSource:0}: Error finding container 19e2eb4a892a44ae780f4aaced1d0727cb7c7e31b881c78918caedaaa97693ed: Status 404 returned error can't find the container with id 19e2eb4a892a44ae780f4aaced1d0727cb7c7e31b881c78918caedaaa97693ed Oct 08 14:18:47 crc kubenswrapper[4789]: W1008 14:18:47.881447 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4509515_2cf4_40a1_9873_588155ad596e.slice/crio-4372a74793ba3a1d74029c158d659c03ff293c78f7baff55f03aae6375ae5748 WatchSource:0}: Error finding container 4372a74793ba3a1d74029c158d659c03ff293c78f7baff55f03aae6375ae5748: Status 404 returned error can't find the container with id 4372a74793ba3a1d74029c158d659c03ff293c78f7baff55f03aae6375ae5748 Oct 08 14:18:47 crc kubenswrapper[4789]: I1008 14:18:47.881681 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 14:18:47 crc kubenswrapper[4789]: I1008 14:18:47.894061 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.008856 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrdhb"] Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.096276 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.178832 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a4509515-2cf4-40a1-9873-588155ad596e","Type":"ContainerStarted","Data":"4372a74793ba3a1d74029c158d659c03ff293c78f7baff55f03aae6375ae5748"} Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.181600 4789 generic.go:334] "Generic (PLEG): container finished" podID="40614905-004f-4a32-89fb-04bc38f13165" containerID="379667279fca355dd18ecc5f603886288081f4394d10272c17070a4c5e6e4348" exitCode=0 Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.181685 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" event={"ID":"40614905-004f-4a32-89fb-04bc38f13165","Type":"ContainerDied","Data":"379667279fca355dd18ecc5f603886288081f4394d10272c17070a4c5e6e4348"} Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.184760 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"32ceb31c-1751-4a04-a91a-b317f3ba3d78","Type":"ContainerStarted","Data":"19e2eb4a892a44ae780f4aaced1d0727cb7c7e31b881c78918caedaaa97693ed"} Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.186606 4789 generic.go:334] "Generic (PLEG): container finished" podID="40054025-fd50-4abc-a3c9-2241ef2be0e9" containerID="783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b" exitCode=0 Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.186689 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" event={"ID":"40054025-fd50-4abc-a3c9-2241ef2be0e9","Type":"ContainerDied","Data":"783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b"} Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.189406 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"05192759-43d6-4ef0-b6f0-6284e2435317","Type":"ContainerStarted","Data":"35ad962c4a320d3afba8876d91185ee116cfee9f18435274a6efaad6fb89f9c7"} Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.191485 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"505cf346-64b1-4de8-8cf6-fca4ddd34cd6","Type":"ContainerStarted","Data":"beb36e4f0acca0233969af488745b3d1ad9b8409d137ddee5d9c750d933b95b2"} Oct 08 14:18:48 crc kubenswrapper[4789]: I1008 14:18:48.239639 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 14:18:49 crc kubenswrapper[4789]: W1008 14:18:49.001576 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaeb67a6_3e80_4d6e_9299_47634a566c72.slice/crio-ffab64301e63b6cb5c0de92fcc0941f43ee102f740a4456bafd75b00b9a5b2a4 WatchSource:0}: Error finding container ffab64301e63b6cb5c0de92fcc0941f43ee102f740a4456bafd75b00b9a5b2a4: Status 404 returned error can't find the container with id ffab64301e63b6cb5c0de92fcc0941f43ee102f740a4456bafd75b00b9a5b2a4 Oct 08 14:18:49 crc kubenswrapper[4789]: W1008 14:18:49.003567 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91ddb1cc_17c9_4f38_ac8a_d78d10798178.slice/crio-f347adc168d51f1cff9288cf8a1931154d5d80fc3bfac082d95d82807dc50604 WatchSource:0}: Error finding container f347adc168d51f1cff9288cf8a1931154d5d80fc3bfac082d95d82807dc50604: Status 404 returned error can't find the container with id f347adc168d51f1cff9288cf8a1931154d5d80fc3bfac082d95d82807dc50604 Oct 08 14:18:49 crc kubenswrapper[4789]: W1008 14:18:49.008339 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1a9507e_04fe_4746_a31b_5ad097c48602.slice/crio-c2500a9b9f9902df62f7f6d6623ec943058cfc71e7416437dad32598e97f2002 WatchSource:0}: Error finding container c2500a9b9f9902df62f7f6d6623ec943058cfc71e7416437dad32598e97f2002: Status 404 returned error can't find the container with id c2500a9b9f9902df62f7f6d6623ec943058cfc71e7416437dad32598e97f2002 Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.015553 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-tkv8v"] Oct 08 14:18:49 crc kubenswrapper[4789]: W1008 14:18:49.018034 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70dc7aa5_b81d_4150_b1dd_57ec365e495c.slice/crio-460e3577ccc54177312f40f7f4edf2199cdfa3991b710eb511775a782132efd9 WatchSource:0}: Error finding container 460e3577ccc54177312f40f7f4edf2199cdfa3991b710eb511775a782132efd9: Status 404 returned error can't find the container with id 460e3577ccc54177312f40f7f4edf2199cdfa3991b710eb511775a782132efd9 Oct 08 14:18:49 crc kubenswrapper[4789]: W1008 14:18:49.041264 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e2285a3_7296_490c_8649_adeae9a0cabc.slice/crio-5661e3c883438171bc311ae714b28fc1a8278a2f87a6e31f012683a071fe32d8 WatchSource:0}: Error finding container 5661e3c883438171bc311ae714b28fc1a8278a2f87a6e31f012683a071fe32d8: Status 404 returned error can't find the container with id 5661e3c883438171bc311ae714b28fc1a8278a2f87a6e31f012683a071fe32d8 Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.096703 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.189735 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.197510 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.202953 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c9336a8d-1117-4da6-a21b-4070f722adc9","Type":"ContainerStarted","Data":"e3f4a50d2a44c2277c7f6e583c4eeef4fadd0de7449e696ac9b577620783eb3f"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.215400 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"55fc0a2c-48c9-4508-ae9d-5350f3298621","Type":"ContainerStarted","Data":"6a761c95276f91455b1d8de776b172e2014ff8e5c08e5130aa1c771e2d35c9ab"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.216867 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx4rp\" (UniqueName: \"kubernetes.io/projected/55261335-33b4-419d-92bc-e5442e647e5d-kube-api-access-nx4rp\") pod \"55261335-33b4-419d-92bc-e5442e647e5d\" (UID: \"55261335-33b4-419d-92bc-e5442e647e5d\") " Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.217125 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55261335-33b4-419d-92bc-e5442e647e5d-config\") pod \"55261335-33b4-419d-92bc-e5442e647e5d\" (UID: \"55261335-33b4-419d-92bc-e5442e647e5d\") " Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.217610 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55261335-33b4-419d-92bc-e5442e647e5d-config" (OuterVolumeSpecName: "config") pod "55261335-33b4-419d-92bc-e5442e647e5d" (UID: "55261335-33b4-419d-92bc-e5442e647e5d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.218034 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"70dc7aa5-b81d-4150-b1dd-57ec365e495c","Type":"ContainerStarted","Data":"460e3577ccc54177312f40f7f4edf2199cdfa3991b710eb511775a782132efd9"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.219667 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrdhb" event={"ID":"91ddb1cc-17c9-4f38-ac8a-d78d10798178","Type":"ContainerStarted","Data":"f347adc168d51f1cff9288cf8a1931154d5d80fc3bfac082d95d82807dc50604"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.221149 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tkv8v" event={"ID":"5e2285a3-7296-490c-8649-adeae9a0cabc","Type":"ContainerStarted","Data":"5661e3c883438171bc311ae714b28fc1a8278a2f87a6e31f012683a071fe32d8"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.222023 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55261335-33b4-419d-92bc-e5442e647e5d-kube-api-access-nx4rp" (OuterVolumeSpecName: "kube-api-access-nx4rp") pod "55261335-33b4-419d-92bc-e5442e647e5d" (UID: "55261335-33b4-419d-92bc-e5442e647e5d"). InnerVolumeSpecName "kube-api-access-nx4rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.225705 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" event={"ID":"12b4276a-e078-439c-a236-0069eadc0a78","Type":"ContainerDied","Data":"e5f655075ab8c52d25f16eb0e865efb8e4fe27a9a3628579f87f7c2bdfe12378"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.225776 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fbc6f7c-m2f76" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.228634 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" event={"ID":"40614905-004f-4a32-89fb-04bc38f13165","Type":"ContainerDied","Data":"7e1ad871f96cf7bb9b29938a38b3a8bde150261e042c2977dc1cf23dbbfc2911"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.228685 4789 scope.go:117] "RemoveContainer" containerID="379667279fca355dd18ecc5f603886288081f4394d10272c17070a4c5e6e4348" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.228824 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d975bffc9-hjkl7" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.234274 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerStarted","Data":"ffab64301e63b6cb5c0de92fcc0941f43ee102f740a4456bafd75b00b9a5b2a4"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.236558 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" event={"ID":"55261335-33b4-419d-92bc-e5442e647e5d","Type":"ContainerDied","Data":"bbb943a5b1eb5af4711a705ac21de6580a068e79e0cbae1e0ee581a748ecfca7"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.236583 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67f9b9c8c7-8glrx" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.239960 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"05192759-43d6-4ef0-b6f0-6284e2435317","Type":"ContainerStarted","Data":"257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.242122 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a1a9507e-04fe-4746-a31b-5ad097c48602","Type":"ContainerStarted","Data":"c2500a9b9f9902df62f7f6d6623ec943058cfc71e7416437dad32598e97f2002"} Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.318520 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-dns-svc\") pod \"40614905-004f-4a32-89fb-04bc38f13165\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.319597 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-dns-svc\") pod \"12b4276a-e078-439c-a236-0069eadc0a78\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.319658 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-config\") pod \"12b4276a-e078-439c-a236-0069eadc0a78\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.319698 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-config\") pod \"40614905-004f-4a32-89fb-04bc38f13165\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.319769 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxgsf\" (UniqueName: \"kubernetes.io/projected/12b4276a-e078-439c-a236-0069eadc0a78-kube-api-access-mxgsf\") pod \"12b4276a-e078-439c-a236-0069eadc0a78\" (UID: \"12b4276a-e078-439c-a236-0069eadc0a78\") " Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.319804 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd6k9\" (UniqueName: \"kubernetes.io/projected/40614905-004f-4a32-89fb-04bc38f13165-kube-api-access-nd6k9\") pod \"40614905-004f-4a32-89fb-04bc38f13165\" (UID: \"40614905-004f-4a32-89fb-04bc38f13165\") " Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.320643 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "12b4276a-e078-439c-a236-0069eadc0a78" (UID: "12b4276a-e078-439c-a236-0069eadc0a78"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.320704 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx4rp\" (UniqueName: \"kubernetes.io/projected/55261335-33b4-419d-92bc-e5442e647e5d-kube-api-access-nx4rp\") on node \"crc\" DevicePath \"\"" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.320722 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55261335-33b4-419d-92bc-e5442e647e5d-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.320807 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-config" (OuterVolumeSpecName: "config") pod "12b4276a-e078-439c-a236-0069eadc0a78" (UID: "12b4276a-e078-439c-a236-0069eadc0a78"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.326645 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40614905-004f-4a32-89fb-04bc38f13165-kube-api-access-nd6k9" (OuterVolumeSpecName: "kube-api-access-nd6k9") pod "40614905-004f-4a32-89fb-04bc38f13165" (UID: "40614905-004f-4a32-89fb-04bc38f13165"). InnerVolumeSpecName "kube-api-access-nd6k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.352315 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67f9b9c8c7-8glrx"] Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.352419 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67f9b9c8c7-8glrx"] Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.359431 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12b4276a-e078-439c-a236-0069eadc0a78-kube-api-access-mxgsf" (OuterVolumeSpecName: "kube-api-access-mxgsf") pod "12b4276a-e078-439c-a236-0069eadc0a78" (UID: "12b4276a-e078-439c-a236-0069eadc0a78"). InnerVolumeSpecName "kube-api-access-mxgsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.380570 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-config" (OuterVolumeSpecName: "config") pod "40614905-004f-4a32-89fb-04bc38f13165" (UID: "40614905-004f-4a32-89fb-04bc38f13165"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.381185 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "40614905-004f-4a32-89fb-04bc38f13165" (UID: "40614905-004f-4a32-89fb-04bc38f13165"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.422199 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.422234 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b4276a-e078-439c-a236-0069eadc0a78-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.422243 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.422253 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxgsf\" (UniqueName: \"kubernetes.io/projected/12b4276a-e078-439c-a236-0069eadc0a78-kube-api-access-mxgsf\") on node \"crc\" DevicePath \"\"" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.422266 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd6k9\" (UniqueName: \"kubernetes.io/projected/40614905-004f-4a32-89fb-04bc38f13165-kube-api-access-nd6k9\") on node \"crc\" DevicePath \"\"" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.422275 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40614905-004f-4a32-89fb-04bc38f13165-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.580589 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fbc6f7c-m2f76"] Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.596743 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8fbc6f7c-m2f76"] Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.610357 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d975bffc9-hjkl7"] Oct 08 14:18:49 crc kubenswrapper[4789]: I1008 14:18:49.614970 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d975bffc9-hjkl7"] Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.256308 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" event={"ID":"40054025-fd50-4abc-a3c9-2241ef2be0e9","Type":"ContainerStarted","Data":"e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f"} Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.257204 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.260398 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"78e021d8-ac60-4092-9e49-bbb3fff88aa7","Type":"ContainerStarted","Data":"8df7d62f5eb0764e8fd662e948f226bceec2d4bdaedbfeb5fe708783e3e5b044"} Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.260778 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.262647 4789 generic.go:334] "Generic (PLEG): container finished" podID="28abfdd3-7f5a-4864-a827-d6c1693ee243" containerID="7d10b08c634972fa218dbdf637e4d5e27061fe28272494225a62a2c89f0eed51" exitCode=0 Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.262697 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68988576c-9hdxg" event={"ID":"28abfdd3-7f5a-4864-a827-d6c1693ee243","Type":"ContainerDied","Data":"7d10b08c634972fa218dbdf637e4d5e27061fe28272494225a62a2c89f0eed51"} Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.276425 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" podStartSLOduration=4.533611069 podStartE2EDuration="23.276409442s" podCreationTimestamp="2025-10-08 14:18:27 +0000 UTC" firstStartedPulling="2025-10-08 14:18:28.667581614 +0000 UTC m=+1048.574329106" lastFinishedPulling="2025-10-08 14:18:47.410379977 +0000 UTC m=+1067.317127479" observedRunningTime="2025-10-08 14:18:50.275070366 +0000 UTC m=+1070.181817858" watchObservedRunningTime="2025-10-08 14:18:50.276409442 +0000 UTC m=+1070.183156934" Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.311286 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=16.378961487 podStartE2EDuration="18.311242736s" podCreationTimestamp="2025-10-08 14:18:32 +0000 UTC" firstStartedPulling="2025-10-08 14:18:47.160269298 +0000 UTC m=+1067.067016790" lastFinishedPulling="2025-10-08 14:18:49.092550547 +0000 UTC m=+1068.999298039" observedRunningTime="2025-10-08 14:18:50.301638276 +0000 UTC m=+1070.208385768" watchObservedRunningTime="2025-10-08 14:18:50.311242736 +0000 UTC m=+1070.217990228" Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.745184 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12b4276a-e078-439c-a236-0069eadc0a78" path="/var/lib/kubelet/pods/12b4276a-e078-439c-a236-0069eadc0a78/volumes" Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.745668 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40614905-004f-4a32-89fb-04bc38f13165" path="/var/lib/kubelet/pods/40614905-004f-4a32-89fb-04bc38f13165/volumes" Oct 08 14:18:50 crc kubenswrapper[4789]: I1008 14:18:50.746370 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55261335-33b4-419d-92bc-e5442e647e5d" path="/var/lib/kubelet/pods/55261335-33b4-419d-92bc-e5442e647e5d/volumes" Oct 08 14:18:54 crc kubenswrapper[4789]: I1008 14:18:54.303976 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68988576c-9hdxg" event={"ID":"28abfdd3-7f5a-4864-a827-d6c1693ee243","Type":"ContainerStarted","Data":"873c7b710294a4b07252cce7b92a0f20d46ce0f52f52cf0ab47cc1079fb0d445"} Oct 08 14:18:54 crc kubenswrapper[4789]: I1008 14:18:54.304684 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:18:54 crc kubenswrapper[4789]: I1008 14:18:54.325723 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68988576c-9hdxg" podStartSLOduration=-9223372009.529074 podStartE2EDuration="27.325702634s" podCreationTimestamp="2025-10-08 14:18:27 +0000 UTC" firstStartedPulling="2025-10-08 14:18:28.223118718 +0000 UTC m=+1048.129866210" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:18:54.32002295 +0000 UTC m=+1074.226770452" watchObservedRunningTime="2025-10-08 14:18:54.325702634 +0000 UTC m=+1074.232450126" Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.320791 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tkv8v" event={"ID":"5e2285a3-7296-490c-8649-adeae9a0cabc","Type":"ContainerStarted","Data":"0fad549eae1cdf7bc3d9bb2b6f37ec75ffe51dcae989be2055b0df29d376dc34"} Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.322330 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"32ceb31c-1751-4a04-a91a-b317f3ba3d78","Type":"ContainerStarted","Data":"9ce80cef0a84e0c98e051b4886a214309c659252b219f2658e30ca2ecab2ddd6"} Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.323709 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a1a9507e-04fe-4746-a31b-5ad097c48602","Type":"ContainerStarted","Data":"84397b21cd289438b85fbeb826462e2b0ca17b9923fa9b7a24b2f9c3c4b0cfcb"} Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.325416 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"70dc7aa5-b81d-4150-b1dd-57ec365e495c","Type":"ContainerStarted","Data":"36b521acd6abf4915e2bbd89eb4140668de4f3c9d7c6d6151ab77b2d4e75a6fe"} Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.326980 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"505cf346-64b1-4de8-8cf6-fca4ddd34cd6","Type":"ContainerStarted","Data":"102da5d689ef84d3c85bf55d0e05b4ce2ba2e105185313d8f1c76f2c935952de"} Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.329668 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrdhb" event={"ID":"91ddb1cc-17c9-4f38-ac8a-d78d10798178","Type":"ContainerStarted","Data":"d189266729dca9adaeaffa64a919d722bec66f7b0af8c90095012f8dc9876a8e"} Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.330589 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-zrdhb" Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.331966 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a4509515-2cf4-40a1-9873-588155ad596e","Type":"ContainerStarted","Data":"459318a2a91161308a2cbf8b5312ea5c90022b0c88f09a77a38c25c7ac3506a8"} Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.332590 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.365668 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zrdhb" podStartSLOduration=14.161778865 podStartE2EDuration="19.36564347s" podCreationTimestamp="2025-10-08 14:18:37 +0000 UTC" firstStartedPulling="2025-10-08 14:18:49.008039036 +0000 UTC m=+1068.914786528" lastFinishedPulling="2025-10-08 14:18:54.211903641 +0000 UTC m=+1074.118651133" observedRunningTime="2025-10-08 14:18:56.365060295 +0000 UTC m=+1076.271807807" watchObservedRunningTime="2025-10-08 14:18:56.36564347 +0000 UTC m=+1076.272390952" Oct 08 14:18:56 crc kubenswrapper[4789]: I1008 14:18:56.400781 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.136371094 podStartE2EDuration="22.400760532s" podCreationTimestamp="2025-10-08 14:18:34 +0000 UTC" firstStartedPulling="2025-10-08 14:18:47.885923015 +0000 UTC m=+1067.792670507" lastFinishedPulling="2025-10-08 14:18:55.150312433 +0000 UTC m=+1075.057059945" observedRunningTime="2025-10-08 14:18:56.39808112 +0000 UTC m=+1076.304828632" watchObservedRunningTime="2025-10-08 14:18:56.400760532 +0000 UTC m=+1076.307508024" Oct 08 14:18:57 crc kubenswrapper[4789]: I1008 14:18:57.350298 4789 generic.go:334] "Generic (PLEG): container finished" podID="5e2285a3-7296-490c-8649-adeae9a0cabc" containerID="0fad549eae1cdf7bc3d9bb2b6f37ec75ffe51dcae989be2055b0df29d376dc34" exitCode=0 Oct 08 14:18:57 crc kubenswrapper[4789]: I1008 14:18:57.351098 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tkv8v" event={"ID":"5e2285a3-7296-490c-8649-adeae9a0cabc","Type":"ContainerDied","Data":"0fad549eae1cdf7bc3d9bb2b6f37ec75ffe51dcae989be2055b0df29d376dc34"} Oct 08 14:18:57 crc kubenswrapper[4789]: I1008 14:18:57.532281 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 08 14:18:58 crc kubenswrapper[4789]: I1008 14:18:58.358908 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerStarted","Data":"468c5cd4b0553b344bb90ce3298ca311aa563ba304e0d98f2b97be854f4efbc1"} Oct 08 14:18:58 crc kubenswrapper[4789]: I1008 14:18:58.362675 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tkv8v" event={"ID":"5e2285a3-7296-490c-8649-adeae9a0cabc","Type":"ContainerStarted","Data":"a2c673793bd3cadc3eed521285e99f62497d16c8a7d6c52be47a34f17f3e2752"} Oct 08 14:18:58 crc kubenswrapper[4789]: I1008 14:18:58.362721 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tkv8v" event={"ID":"5e2285a3-7296-490c-8649-adeae9a0cabc","Type":"ContainerStarted","Data":"4e393d3b314e2c3ea383829d22f59c92fe7704ebf3ce7d5bf9bf57d8d048575a"} Oct 08 14:18:58 crc kubenswrapper[4789]: I1008 14:18:58.363611 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:58 crc kubenswrapper[4789]: I1008 14:18:58.363718 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:18:58 crc kubenswrapper[4789]: I1008 14:18:58.388355 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:18:58 crc kubenswrapper[4789]: I1008 14:18:58.411502 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-tkv8v" podStartSLOduration=16.49132069 podStartE2EDuration="21.411483186s" podCreationTimestamp="2025-10-08 14:18:37 +0000 UTC" firstStartedPulling="2025-10-08 14:18:49.08123656 +0000 UTC m=+1068.987984052" lastFinishedPulling="2025-10-08 14:18:54.001399056 +0000 UTC m=+1073.908146548" observedRunningTime="2025-10-08 14:18:58.410647793 +0000 UTC m=+1078.317395285" watchObservedRunningTime="2025-10-08 14:18:58.411483186 +0000 UTC m=+1078.318230668" Oct 08 14:18:58 crc kubenswrapper[4789]: I1008 14:18:58.471580 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68988576c-9hdxg"] Oct 08 14:18:58 crc kubenswrapper[4789]: I1008 14:18:58.471852 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68988576c-9hdxg" podUID="28abfdd3-7f5a-4864-a827-d6c1693ee243" containerName="dnsmasq-dns" containerID="cri-o://873c7b710294a4b07252cce7b92a0f20d46ce0f52f52cf0ab47cc1079fb0d445" gracePeriod=10 Oct 08 14:18:59 crc kubenswrapper[4789]: I1008 14:18:59.374740 4789 generic.go:334] "Generic (PLEG): container finished" podID="28abfdd3-7f5a-4864-a827-d6c1693ee243" containerID="873c7b710294a4b07252cce7b92a0f20d46ce0f52f52cf0ab47cc1079fb0d445" exitCode=0 Oct 08 14:18:59 crc kubenswrapper[4789]: I1008 14:18:59.374804 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68988576c-9hdxg" event={"ID":"28abfdd3-7f5a-4864-a827-d6c1693ee243","Type":"ContainerDied","Data":"873c7b710294a4b07252cce7b92a0f20d46ce0f52f52cf0ab47cc1079fb0d445"} Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.063818 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.226523 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsp6k\" (UniqueName: \"kubernetes.io/projected/28abfdd3-7f5a-4864-a827-d6c1693ee243-kube-api-access-gsp6k\") pod \"28abfdd3-7f5a-4864-a827-d6c1693ee243\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.226582 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-dns-svc\") pod \"28abfdd3-7f5a-4864-a827-d6c1693ee243\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.226628 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-config\") pod \"28abfdd3-7f5a-4864-a827-d6c1693ee243\" (UID: \"28abfdd3-7f5a-4864-a827-d6c1693ee243\") " Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.230556 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28abfdd3-7f5a-4864-a827-d6c1693ee243-kube-api-access-gsp6k" (OuterVolumeSpecName: "kube-api-access-gsp6k") pod "28abfdd3-7f5a-4864-a827-d6c1693ee243" (UID: "28abfdd3-7f5a-4864-a827-d6c1693ee243"). InnerVolumeSpecName "kube-api-access-gsp6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.277748 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "28abfdd3-7f5a-4864-a827-d6c1693ee243" (UID: "28abfdd3-7f5a-4864-a827-d6c1693ee243"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.328351 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsp6k\" (UniqueName: \"kubernetes.io/projected/28abfdd3-7f5a-4864-a827-d6c1693ee243-kube-api-access-gsp6k\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.328390 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.353525 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-config" (OuterVolumeSpecName: "config") pod "28abfdd3-7f5a-4864-a827-d6c1693ee243" (UID: "28abfdd3-7f5a-4864-a827-d6c1693ee243"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.398623 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68988576c-9hdxg" event={"ID":"28abfdd3-7f5a-4864-a827-d6c1693ee243","Type":"ContainerDied","Data":"3f2e799a9f7f443ab6c5744c38935b8fbc87b75053280e8be4ae41da87db0cd1"} Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.398910 4789 scope.go:117] "RemoveContainer" containerID="873c7b710294a4b07252cce7b92a0f20d46ce0f52f52cf0ab47cc1079fb0d445" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.398711 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68988576c-9hdxg" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.427133 4789 scope.go:117] "RemoveContainer" containerID="7d10b08c634972fa218dbdf637e4d5e27061fe28272494225a62a2c89f0eed51" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.430122 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28abfdd3-7f5a-4864-a827-d6c1693ee243-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.433186 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68988576c-9hdxg"] Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.443154 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68988576c-9hdxg"] Oct 08 14:19:00 crc kubenswrapper[4789]: I1008 14:19:00.748705 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28abfdd3-7f5a-4864-a827-d6c1693ee243" path="/var/lib/kubelet/pods/28abfdd3-7f5a-4864-a827-d6c1693ee243/volumes" Oct 08 14:19:01 crc kubenswrapper[4789]: I1008 14:19:01.408570 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a1a9507e-04fe-4746-a31b-5ad097c48602","Type":"ContainerStarted","Data":"62067c8f51e33adbf9f21779444756f31afef292195faa6043f82513adedbd87"} Oct 08 14:19:01 crc kubenswrapper[4789]: I1008 14:19:01.410399 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"70dc7aa5-b81d-4150-b1dd-57ec365e495c","Type":"ContainerStarted","Data":"2e0d1ef0c0c19b2842a18bd2c83451e9800ffb2ca086b0dea9fdc62b26f5781a"} Oct 08 14:19:01 crc kubenswrapper[4789]: I1008 14:19:01.437123 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=10.361911018 podStartE2EDuration="21.437103645s" podCreationTimestamp="2025-10-08 14:18:40 +0000 UTC" firstStartedPulling="2025-10-08 14:18:49.024194394 +0000 UTC m=+1068.930941886" lastFinishedPulling="2025-10-08 14:19:00.099387021 +0000 UTC m=+1080.006134513" observedRunningTime="2025-10-08 14:19:01.432041968 +0000 UTC m=+1081.338789460" watchObservedRunningTime="2025-10-08 14:19:01.437103645 +0000 UTC m=+1081.343851147" Oct 08 14:19:02 crc kubenswrapper[4789]: I1008 14:19:02.148831 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.149014 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.203739 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.480787 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.754549 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5677ffc7f-p8lx8"] Oct 08 14:19:03 crc kubenswrapper[4789]: E1008 14:19:03.755817 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40614905-004f-4a32-89fb-04bc38f13165" containerName="init" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.756045 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="40614905-004f-4a32-89fb-04bc38f13165" containerName="init" Oct 08 14:19:03 crc kubenswrapper[4789]: E1008 14:19:03.756224 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28abfdd3-7f5a-4864-a827-d6c1693ee243" containerName="dnsmasq-dns" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.756315 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="28abfdd3-7f5a-4864-a827-d6c1693ee243" containerName="dnsmasq-dns" Oct 08 14:19:03 crc kubenswrapper[4789]: E1008 14:19:03.756414 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28abfdd3-7f5a-4864-a827-d6c1693ee243" containerName="init" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.756506 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="28abfdd3-7f5a-4864-a827-d6c1693ee243" containerName="init" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.756754 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="28abfdd3-7f5a-4864-a827-d6c1693ee243" containerName="dnsmasq-dns" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.756854 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="40614905-004f-4a32-89fb-04bc38f13165" containerName="init" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.758097 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.764058 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.770723 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5677ffc7f-p8lx8"] Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.873428 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-nrhvq"] Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.874476 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.879299 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.879563 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nrhvq"] Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.887921 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-config\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.888274 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-ovsdbserver-sb\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.888381 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sznk\" (UniqueName: \"kubernetes.io/projected/141f2b14-7eb3-4345-b5ee-ee9666712853-kube-api-access-6sznk\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.888464 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-dns-svc\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990054 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-combined-ca-bundle\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990108 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-config\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990125 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-ovn-rundir\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990161 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-config\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990193 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-ovsdbserver-sb\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990224 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990259 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sznk\" (UniqueName: \"kubernetes.io/projected/141f2b14-7eb3-4345-b5ee-ee9666712853-kube-api-access-6sznk\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990295 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-ovs-rundir\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990315 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-dns-svc\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.990334 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvmmh\" (UniqueName: \"kubernetes.io/projected/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-kube-api-access-pvmmh\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.991091 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-config\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.991230 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-ovsdbserver-sb\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:03 crc kubenswrapper[4789]: I1008 14:19:03.991328 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-dns-svc\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.007730 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sznk\" (UniqueName: \"kubernetes.io/projected/141f2b14-7eb3-4345-b5ee-ee9666712853-kube-api-access-6sznk\") pod \"dnsmasq-dns-5677ffc7f-p8lx8\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.079266 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.091532 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-config\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.091628 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.091694 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-ovs-rundir\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.091730 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvmmh\" (UniqueName: \"kubernetes.io/projected/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-kube-api-access-pvmmh\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.091795 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-combined-ca-bundle\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.091817 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-ovn-rundir\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.092103 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-ovn-rundir\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.092095 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-ovs-rundir\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.092441 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-config\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.101719 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-combined-ca-bundle\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.105484 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.119764 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvmmh\" (UniqueName: \"kubernetes.io/projected/23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718-kube-api-access-pvmmh\") pod \"ovn-controller-metrics-nrhvq\" (UID: \"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718\") " pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.191793 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5677ffc7f-p8lx8"] Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.196338 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nrhvq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.233392 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f6954cb7f-4mjzq"] Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.235104 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.239124 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.249619 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f6954cb7f-4mjzq"] Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.397982 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-dns-svc\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.398464 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-config\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.398551 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.398778 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/a7d4a0e7-d24d-4381-83bc-23b094382b0b-kube-api-access-6jz5f\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.398923 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.503062 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-dns-svc\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.503159 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-config\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.503283 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.503353 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/a7d4a0e7-d24d-4381-83bc-23b094382b0b-kube-api-access-6jz5f\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.503417 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.504064 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-dns-svc\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.504451 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.504804 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.505002 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-config\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.536953 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/a7d4a0e7-d24d-4381-83bc-23b094382b0b-kube-api-access-6jz5f\") pod \"dnsmasq-dns-5f6954cb7f-4mjzq\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.616711 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nrhvq"] Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.657128 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.770040 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5677ffc7f-p8lx8"] Oct 08 14:19:04 crc kubenswrapper[4789]: I1008 14:19:04.808134 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.107713 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f6954cb7f-4mjzq"] Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.159150 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58f55f8967-2cng7"] Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.160650 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.172715 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58f55f8967-2cng7"] Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.236654 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-nb\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.236733 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-config\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.236788 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-dns-svc\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.236850 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbvrg\" (UniqueName: \"kubernetes.io/projected/1b0c0d62-4dc5-49af-bd91-746237228359-kube-api-access-jbvrg\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.236892 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-sb\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.337913 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbvrg\" (UniqueName: \"kubernetes.io/projected/1b0c0d62-4dc5-49af-bd91-746237228359-kube-api-access-jbvrg\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.338007 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-sb\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.338079 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-nb\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.338116 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-config\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.338164 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-dns-svc\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.339465 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-config\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.339479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-sb\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.340527 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-nb\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.340605 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-dns-svc\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.368927 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbvrg\" (UniqueName: \"kubernetes.io/projected/1b0c0d62-4dc5-49af-bd91-746237228359-kube-api-access-jbvrg\") pod \"dnsmasq-dns-58f55f8967-2cng7\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.452735 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nrhvq" event={"ID":"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718","Type":"ContainerStarted","Data":"630bad023b5d171df410dfcf84a3c86f726149bd6ba686a89621c88dc03c22a7"} Oct 08 14:19:05 crc kubenswrapper[4789]: W1008 14:19:05.453665 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7d4a0e7_d24d_4381_83bc_23b094382b0b.slice/crio-d2b7b1884fe2df817761659d4be64376fc54c12e75c760a6ec591c45c7cecbf1 WatchSource:0}: Error finding container d2b7b1884fe2df817761659d4be64376fc54c12e75c760a6ec591c45c7cecbf1: Status 404 returned error can't find the container with id d2b7b1884fe2df817761659d4be64376fc54c12e75c760a6ec591c45c7cecbf1 Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.454375 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" event={"ID":"141f2b14-7eb3-4345-b5ee-ee9666712853","Type":"ContainerStarted","Data":"e6a306cb4b1089d254b29a4fbd7c7a910ae1637a913e8a2bb08cd813d40e4d51"} Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.456562 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f6954cb7f-4mjzq"] Oct 08 14:19:05 crc kubenswrapper[4789]: I1008 14:19:05.491422 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.003527 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58f55f8967-2cng7"] Oct 08 14:19:06 crc kubenswrapper[4789]: W1008 14:19:06.007239 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b0c0d62_4dc5_49af_bd91_746237228359.slice/crio-d3fa09c60e4b27d7623f47097e723bae79dc17330c28ccfbd26bf515d0d61548 WatchSource:0}: Error finding container d3fa09c60e4b27d7623f47097e723bae79dc17330c28ccfbd26bf515d0d61548: Status 404 returned error can't find the container with id d3fa09c60e4b27d7623f47097e723bae79dc17330c28ccfbd26bf515d0d61548 Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.267232 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.273267 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.276741 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-2p6g8" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.276811 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.280723 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.280969 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.287023 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.360558 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.360616 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0462c12d-8df3-4734-9c2e-a925c9c06b9e-lock\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.360638 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.360775 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wr5w\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-kube-api-access-4wr5w\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.360888 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0462c12d-8df3-4734-9c2e-a925c9c06b9e-cache\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.461684 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" event={"ID":"1b0c0d62-4dc5-49af-bd91-746237228359","Type":"ContainerStarted","Data":"d3fa09c60e4b27d7623f47097e723bae79dc17330c28ccfbd26bf515d0d61548"} Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.462584 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.462651 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0462c12d-8df3-4734-9c2e-a925c9c06b9e-lock\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.462673 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: E1008 14:19:06.462783 4789 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 14:19:06 crc kubenswrapper[4789]: E1008 14:19:06.462815 4789 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 14:19:06 crc kubenswrapper[4789]: E1008 14:19:06.462859 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift podName:0462c12d-8df3-4734-9c2e-a925c9c06b9e nodeName:}" failed. No retries permitted until 2025-10-08 14:19:06.962842875 +0000 UTC m=+1086.869590367 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift") pod "swift-storage-0" (UID: "0462c12d-8df3-4734-9c2e-a925c9c06b9e") : configmap "swift-ring-files" not found Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.463049 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" event={"ID":"a7d4a0e7-d24d-4381-83bc-23b094382b0b","Type":"ContainerStarted","Data":"d2b7b1884fe2df817761659d4be64376fc54c12e75c760a6ec591c45c7cecbf1"} Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.463079 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0462c12d-8df3-4734-9c2e-a925c9c06b9e-lock\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.463114 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wr5w\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-kube-api-access-4wr5w\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.463276 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.463832 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0462c12d-8df3-4734-9c2e-a925c9c06b9e-cache\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.464320 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0462c12d-8df3-4734-9c2e-a925c9c06b9e-cache\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.483742 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wr5w\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-kube-api-access-4wr5w\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.490314 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.788638 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=19.691811101 podStartE2EDuration="30.788613004s" podCreationTimestamp="2025-10-08 14:18:36 +0000 UTC" firstStartedPulling="2025-10-08 14:18:49.011705646 +0000 UTC m=+1068.918453138" lastFinishedPulling="2025-10-08 14:19:00.108507549 +0000 UTC m=+1080.015255041" observedRunningTime="2025-10-08 14:19:06.49033084 +0000 UTC m=+1086.397078332" watchObservedRunningTime="2025-10-08 14:19:06.788613004 +0000 UTC m=+1086.695360496" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.793496 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-w7mzz"] Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.794885 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.799046 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.799298 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.803610 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.826067 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-w7mzz"] Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.867482 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-w7mzz"] Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.872473 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-ring-data-devices\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.872526 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-scripts\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.872551 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-dispersionconf\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.872655 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-combined-ca-bundle\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.872681 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-swiftconf\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.872762 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vwgh\" (UniqueName: \"kubernetes.io/projected/840c0144-c547-42f3-a6ed-a85917a10f98-kube-api-access-2vwgh\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.872810 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/840c0144-c547-42f3-a6ed-a85917a10f98-etc-swift\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: E1008 14:19:06.904783 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-2vwgh ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-w7mzz" podUID="840c0144-c547-42f3-a6ed-a85917a10f98" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.915045 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-pr6sr"] Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.916446 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.972075 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-pr6sr"] Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.973927 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-scripts\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974031 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-ring-data-devices\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974070 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-scripts\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974093 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-dispersionconf\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974155 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-swiftconf\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974184 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7b17c27-5518-480a-875f-091ae1dd3606-etc-swift\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974206 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-combined-ca-bundle\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974243 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974271 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-combined-ca-bundle\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974299 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-swiftconf\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974335 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgv4r\" (UniqueName: \"kubernetes.io/projected/a7b17c27-5518-480a-875f-091ae1dd3606-kube-api-access-qgv4r\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974356 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-ring-data-devices\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974396 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-dispersionconf\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974429 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vwgh\" (UniqueName: \"kubernetes.io/projected/840c0144-c547-42f3-a6ed-a85917a10f98-kube-api-access-2vwgh\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974465 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/840c0144-c547-42f3-a6ed-a85917a10f98-etc-swift\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.974923 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/840c0144-c547-42f3-a6ed-a85917a10f98-etc-swift\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: E1008 14:19:06.975090 4789 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 14:19:06 crc kubenswrapper[4789]: E1008 14:19:06.975119 4789 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 14:19:06 crc kubenswrapper[4789]: E1008 14:19:06.975174 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift podName:0462c12d-8df3-4734-9c2e-a925c9c06b9e nodeName:}" failed. No retries permitted until 2025-10-08 14:19:07.975151989 +0000 UTC m=+1087.881899491 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift") pod "swift-storage-0" (UID: "0462c12d-8df3-4734-9c2e-a925c9c06b9e") : configmap "swift-ring-files" not found Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.975274 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-ring-data-devices\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.975778 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-scripts\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.993373 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-dispersionconf\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.994042 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-swiftconf\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:06 crc kubenswrapper[4789]: I1008 14:19:06.997732 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-combined-ca-bundle\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.013724 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vwgh\" (UniqueName: \"kubernetes.io/projected/840c0144-c547-42f3-a6ed-a85917a10f98-kube-api-access-2vwgh\") pod \"swift-ring-rebalance-w7mzz\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.076039 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-swiftconf\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.076103 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7b17c27-5518-480a-875f-091ae1dd3606-etc-swift\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.076120 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-combined-ca-bundle\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.076171 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgv4r\" (UniqueName: \"kubernetes.io/projected/a7b17c27-5518-480a-875f-091ae1dd3606-kube-api-access-qgv4r\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.076189 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-ring-data-devices\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.076226 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-dispersionconf\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.076275 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-scripts\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.076571 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7b17c27-5518-480a-875f-091ae1dd3606-etc-swift\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.076965 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-scripts\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.077286 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-ring-data-devices\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.080045 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-swiftconf\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.080515 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-combined-ca-bundle\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.086379 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-dispersionconf\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.106597 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgv4r\" (UniqueName: \"kubernetes.io/projected/a7b17c27-5518-480a-875f-091ae1dd3606-kube-api-access-qgv4r\") pod \"swift-ring-rebalance-pr6sr\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.238531 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.479510 4789 generic.go:334] "Generic (PLEG): container finished" podID="1b0c0d62-4dc5-49af-bd91-746237228359" containerID="43ca0225ac870080925935c167059e44f176fab94ada40fbe3c6a60fb950fa6a" exitCode=0 Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.480499 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" event={"ID":"1b0c0d62-4dc5-49af-bd91-746237228359","Type":"ContainerDied","Data":"43ca0225ac870080925935c167059e44f176fab94ada40fbe3c6a60fb950fa6a"} Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.483166 4789 generic.go:334] "Generic (PLEG): container finished" podID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerID="468c5cd4b0553b344bb90ce3298ca311aa563ba304e0d98f2b97be854f4efbc1" exitCode=0 Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.483310 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.483301 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerDied","Data":"468c5cd4b0553b344bb90ce3298ca311aa563ba304e0d98f2b97be854f4efbc1"} Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.496525 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.589514 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-scripts" (OuterVolumeSpecName: "scripts") pod "840c0144-c547-42f3-a6ed-a85917a10f98" (UID: "840c0144-c547-42f3-a6ed-a85917a10f98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.589848 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-scripts\") pod \"840c0144-c547-42f3-a6ed-a85917a10f98\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.589949 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-swiftconf\") pod \"840c0144-c547-42f3-a6ed-a85917a10f98\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.590102 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-combined-ca-bundle\") pod \"840c0144-c547-42f3-a6ed-a85917a10f98\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.590479 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-dispersionconf\") pod \"840c0144-c547-42f3-a6ed-a85917a10f98\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.591149 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/840c0144-c547-42f3-a6ed-a85917a10f98-etc-swift\") pod \"840c0144-c547-42f3-a6ed-a85917a10f98\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.591289 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vwgh\" (UniqueName: \"kubernetes.io/projected/840c0144-c547-42f3-a6ed-a85917a10f98-kube-api-access-2vwgh\") pod \"840c0144-c547-42f3-a6ed-a85917a10f98\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.591413 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-ring-data-devices\") pod \"840c0144-c547-42f3-a6ed-a85917a10f98\" (UID: \"840c0144-c547-42f3-a6ed-a85917a10f98\") " Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.591654 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/840c0144-c547-42f3-a6ed-a85917a10f98-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "840c0144-c547-42f3-a6ed-a85917a10f98" (UID: "840c0144-c547-42f3-a6ed-a85917a10f98"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.592269 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "840c0144-c547-42f3-a6ed-a85917a10f98" (UID: "840c0144-c547-42f3-a6ed-a85917a10f98"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.594372 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "840c0144-c547-42f3-a6ed-a85917a10f98" (UID: "840c0144-c547-42f3-a6ed-a85917a10f98"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.594755 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "840c0144-c547-42f3-a6ed-a85917a10f98" (UID: "840c0144-c547-42f3-a6ed-a85917a10f98"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.595264 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "840c0144-c547-42f3-a6ed-a85917a10f98" (UID: "840c0144-c547-42f3-a6ed-a85917a10f98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.595347 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/840c0144-c547-42f3-a6ed-a85917a10f98-kube-api-access-2vwgh" (OuterVolumeSpecName: "kube-api-access-2vwgh") pod "840c0144-c547-42f3-a6ed-a85917a10f98" (UID: "840c0144-c547-42f3-a6ed-a85917a10f98"). InnerVolumeSpecName "kube-api-access-2vwgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.596170 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.596917 4789 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.597036 4789 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/840c0144-c547-42f3-a6ed-a85917a10f98-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.597116 4789 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.597209 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/840c0144-c547-42f3-a6ed-a85917a10f98-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.597279 4789 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/840c0144-c547-42f3-a6ed-a85917a10f98-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.700530 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vwgh\" (UniqueName: \"kubernetes.io/projected/840c0144-c547-42f3-a6ed-a85917a10f98-kube-api-access-2vwgh\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:07 crc kubenswrapper[4789]: I1008 14:19:07.705628 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-pr6sr"] Oct 08 14:19:07 crc kubenswrapper[4789]: W1008 14:19:07.710337 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7b17c27_5518_480a_875f_091ae1dd3606.slice/crio-36fee5c2a6ddff8971e8cd8dd3581e8ae6d06ec6b619ad2528689fe312254fb5 WatchSource:0}: Error finding container 36fee5c2a6ddff8971e8cd8dd3581e8ae6d06ec6b619ad2528689fe312254fb5: Status 404 returned error can't find the container with id 36fee5c2a6ddff8971e8cd8dd3581e8ae6d06ec6b619ad2528689fe312254fb5 Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.011160 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:08 crc kubenswrapper[4789]: E1008 14:19:08.011387 4789 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 14:19:08 crc kubenswrapper[4789]: E1008 14:19:08.011427 4789 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 14:19:08 crc kubenswrapper[4789]: E1008 14:19:08.011489 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift podName:0462c12d-8df3-4734-9c2e-a925c9c06b9e nodeName:}" failed. No retries permitted until 2025-10-08 14:19:10.011466435 +0000 UTC m=+1089.918213927 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift") pod "swift-storage-0" (UID: "0462c12d-8df3-4734-9c2e-a925c9c06b9e") : configmap "swift-ring-files" not found Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.125488 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.125849 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.166277 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.502570 4789 generic.go:334] "Generic (PLEG): container finished" podID="505cf346-64b1-4de8-8cf6-fca4ddd34cd6" containerID="102da5d689ef84d3c85bf55d0e05b4ce2ba2e105185313d8f1c76f2c935952de" exitCode=0 Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.502644 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"505cf346-64b1-4de8-8cf6-fca4ddd34cd6","Type":"ContainerDied","Data":"102da5d689ef84d3c85bf55d0e05b4ce2ba2e105185313d8f1c76f2c935952de"} Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.508307 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" event={"ID":"1b0c0d62-4dc5-49af-bd91-746237228359","Type":"ContainerStarted","Data":"e187ba8bfc6f6962c902db0c44f12c868ae0283bb53c8d1f44a15ca43db752c3"} Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.508435 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.511088 4789 generic.go:334] "Generic (PLEG): container finished" podID="32ceb31c-1751-4a04-a91a-b317f3ba3d78" containerID="9ce80cef0a84e0c98e051b4886a214309c659252b219f2658e30ca2ecab2ddd6" exitCode=0 Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.511162 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"32ceb31c-1751-4a04-a91a-b317f3ba3d78","Type":"ContainerDied","Data":"9ce80cef0a84e0c98e051b4886a214309c659252b219f2658e30ca2ecab2ddd6"} Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.525433 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w7mzz" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.525447 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pr6sr" event={"ID":"a7b17c27-5518-480a-875f-091ae1dd3606","Type":"ContainerStarted","Data":"36fee5c2a6ddff8971e8cd8dd3581e8ae6d06ec6b619ad2528689fe312254fb5"} Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.586912 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.590162 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" podStartSLOduration=3.590143928 podStartE2EDuration="3.590143928s" podCreationTimestamp="2025-10-08 14:19:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:08.585089251 +0000 UTC m=+1088.491836753" watchObservedRunningTime="2025-10-08 14:19:08.590143928 +0000 UTC m=+1088.496891420" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.753382 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.755417 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.763113 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.763316 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-zm5hc" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.763338 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.763721 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.768364 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.781862 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-w7mzz"] Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.794167 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-w7mzz"] Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.833203 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd23ae36-37d1-458e-ab17-0f0db57e57d3-config\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.833244 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd23ae36-37d1-458e-ab17-0f0db57e57d3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.833308 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd23ae36-37d1-458e-ab17-0f0db57e57d3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.833391 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd23ae36-37d1-458e-ab17-0f0db57e57d3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.833413 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd23ae36-37d1-458e-ab17-0f0db57e57d3-scripts\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.833450 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtm5j\" (UniqueName: \"kubernetes.io/projected/fd23ae36-37d1-458e-ab17-0f0db57e57d3-kube-api-access-dtm5j\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.833467 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd23ae36-37d1-458e-ab17-0f0db57e57d3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.935020 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd23ae36-37d1-458e-ab17-0f0db57e57d3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.935065 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd23ae36-37d1-458e-ab17-0f0db57e57d3-scripts\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.935189 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtm5j\" (UniqueName: \"kubernetes.io/projected/fd23ae36-37d1-458e-ab17-0f0db57e57d3-kube-api-access-dtm5j\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.935224 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd23ae36-37d1-458e-ab17-0f0db57e57d3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.935857 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd23ae36-37d1-458e-ab17-0f0db57e57d3-scripts\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.936177 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd23ae36-37d1-458e-ab17-0f0db57e57d3-config\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.936204 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd23ae36-37d1-458e-ab17-0f0db57e57d3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.936257 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd23ae36-37d1-458e-ab17-0f0db57e57d3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.936970 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd23ae36-37d1-458e-ab17-0f0db57e57d3-config\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.938196 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd23ae36-37d1-458e-ab17-0f0db57e57d3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.940328 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd23ae36-37d1-458e-ab17-0f0db57e57d3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.941726 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd23ae36-37d1-458e-ab17-0f0db57e57d3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.942162 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd23ae36-37d1-458e-ab17-0f0db57e57d3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:08 crc kubenswrapper[4789]: I1008 14:19:08.956139 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtm5j\" (UniqueName: \"kubernetes.io/projected/fd23ae36-37d1-458e-ab17-0f0db57e57d3-kube-api-access-dtm5j\") pod \"ovn-northd-0\" (UID: \"fd23ae36-37d1-458e-ab17-0f0db57e57d3\") " pod="openstack/ovn-northd-0" Oct 08 14:19:09 crc kubenswrapper[4789]: I1008 14:19:09.113522 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 14:19:09 crc kubenswrapper[4789]: I1008 14:19:09.536715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"32ceb31c-1751-4a04-a91a-b317f3ba3d78","Type":"ContainerStarted","Data":"11b877cd69fca0d4ca743f3b7cbe5dceff00ce8458043e017400b19e03752bca"} Oct 08 14:19:09 crc kubenswrapper[4789]: I1008 14:19:09.539766 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"505cf346-64b1-4de8-8cf6-fca4ddd34cd6","Type":"ContainerStarted","Data":"6841819ae690451e3ccdcdf15a9570e954a9c9c226a9d283c07589c5e9b193e4"} Oct 08 14:19:09 crc kubenswrapper[4789]: I1008 14:19:09.573038 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 14:19:09 crc kubenswrapper[4789]: I1008 14:19:09.622189 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=33.368448761 podStartE2EDuration="39.609553326s" podCreationTimestamp="2025-10-08 14:18:30 +0000 UTC" firstStartedPulling="2025-10-08 14:18:47.760095135 +0000 UTC m=+1067.666842627" lastFinishedPulling="2025-10-08 14:18:54.0011997 +0000 UTC m=+1073.907947192" observedRunningTime="2025-10-08 14:19:09.584663951 +0000 UTC m=+1089.491411443" watchObservedRunningTime="2025-10-08 14:19:09.609553326 +0000 UTC m=+1089.516300818" Oct 08 14:19:10 crc kubenswrapper[4789]: I1008 14:19:10.071659 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:10 crc kubenswrapper[4789]: E1008 14:19:10.071911 4789 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 14:19:10 crc kubenswrapper[4789]: E1008 14:19:10.071925 4789 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 14:19:10 crc kubenswrapper[4789]: E1008 14:19:10.071966 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift podName:0462c12d-8df3-4734-9c2e-a925c9c06b9e nodeName:}" failed. No retries permitted until 2025-10-08 14:19:14.071952758 +0000 UTC m=+1093.978700250 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift") pod "swift-storage-0" (UID: "0462c12d-8df3-4734-9c2e-a925c9c06b9e") : configmap "swift-ring-files" not found Oct 08 14:19:10 crc kubenswrapper[4789]: I1008 14:19:10.551594 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fd23ae36-37d1-458e-ab17-0f0db57e57d3","Type":"ContainerStarted","Data":"aab998eee0bb81b9e0d3fda6d27f326c439c5ce959e0f10e43e1d650346067e9"} Oct 08 14:19:10 crc kubenswrapper[4789]: I1008 14:19:10.578207 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=33.966985572 podStartE2EDuration="40.578186728s" podCreationTimestamp="2025-10-08 14:18:30 +0000 UTC" firstStartedPulling="2025-10-08 14:18:47.238596261 +0000 UTC m=+1067.145343753" lastFinishedPulling="2025-10-08 14:18:53.849797407 +0000 UTC m=+1073.756544909" observedRunningTime="2025-10-08 14:19:10.575254668 +0000 UTC m=+1090.482002180" watchObservedRunningTime="2025-10-08 14:19:10.578186728 +0000 UTC m=+1090.484934220" Oct 08 14:19:10 crc kubenswrapper[4789]: I1008 14:19:10.740611 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="840c0144-c547-42f3-a6ed-a85917a10f98" path="/var/lib/kubelet/pods/840c0144-c547-42f3-a6ed-a85917a10f98/volumes" Oct 08 14:19:12 crc kubenswrapper[4789]: I1008 14:19:12.037553 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 08 14:19:12 crc kubenswrapper[4789]: I1008 14:19:12.037619 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 08 14:19:12 crc kubenswrapper[4789]: I1008 14:19:12.057737 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 08 14:19:12 crc kubenswrapper[4789]: I1008 14:19:12.057807 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 08 14:19:14 crc kubenswrapper[4789]: I1008 14:19:14.139423 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:14 crc kubenswrapper[4789]: E1008 14:19:14.139585 4789 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 14:19:14 crc kubenswrapper[4789]: E1008 14:19:14.139752 4789 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 14:19:14 crc kubenswrapper[4789]: E1008 14:19:14.139805 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift podName:0462c12d-8df3-4734-9c2e-a925c9c06b9e nodeName:}" failed. No retries permitted until 2025-10-08 14:19:22.139789034 +0000 UTC m=+1102.046536526 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift") pod "swift-storage-0" (UID: "0462c12d-8df3-4734-9c2e-a925c9c06b9e") : configmap "swift-ring-files" not found Oct 08 14:19:15 crc kubenswrapper[4789]: I1008 14:19:15.493914 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:15 crc kubenswrapper[4789]: I1008 14:19:15.556866 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7547999-h5qmx"] Oct 08 14:19:15 crc kubenswrapper[4789]: I1008 14:19:15.557136 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" podUID="40054025-fd50-4abc-a3c9-2241ef2be0e9" containerName="dnsmasq-dns" containerID="cri-o://e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f" gracePeriod=10 Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.524179 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.615971 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nrhvq" event={"ID":"23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718","Type":"ContainerStarted","Data":"e6eee04728d6ca1b68e3bba5affedc26a3332a90bdf5a71e718ff44d64651805"} Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.618146 4789 generic.go:334] "Generic (PLEG): container finished" podID="a7d4a0e7-d24d-4381-83bc-23b094382b0b" containerID="07f0cdfd9c0b5c8ad03a5ffbe321024b8bbf36542fc5fb9d3f6a416e1859e5e9" exitCode=0 Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.618207 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" event={"ID":"a7d4a0e7-d24d-4381-83bc-23b094382b0b","Type":"ContainerDied","Data":"07f0cdfd9c0b5c8ad03a5ffbe321024b8bbf36542fc5fb9d3f6a416e1859e5e9"} Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.620546 4789 generic.go:334] "Generic (PLEG): container finished" podID="141f2b14-7eb3-4345-b5ee-ee9666712853" containerID="88a1e2db7673690b02ff31d512303bbee22976fb2635059c8cb0a0c558f1ac89" exitCode=0 Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.620588 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" event={"ID":"141f2b14-7eb3-4345-b5ee-ee9666712853","Type":"ContainerDied","Data":"88a1e2db7673690b02ff31d512303bbee22976fb2635059c8cb0a0c558f1ac89"} Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.624414 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fd23ae36-37d1-458e-ab17-0f0db57e57d3","Type":"ContainerStarted","Data":"191b78b00b64039196c42c04e087974970a6eeedf8f6c4647aaf679e2749bf87"} Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.631107 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pr6sr" event={"ID":"a7b17c27-5518-480a-875f-091ae1dd3606","Type":"ContainerStarted","Data":"ab0b023a33b800df6b2d7de7569d43ca0fa1aef5e07dee2962ca4acbea9445eb"} Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.636692 4789 generic.go:334] "Generic (PLEG): container finished" podID="40054025-fd50-4abc-a3c9-2241ef2be0e9" containerID="e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f" exitCode=0 Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.636733 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" event={"ID":"40054025-fd50-4abc-a3c9-2241ef2be0e9","Type":"ContainerDied","Data":"e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f"} Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.636757 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" event={"ID":"40054025-fd50-4abc-a3c9-2241ef2be0e9","Type":"ContainerDied","Data":"4deea8a8c2275f9660fbb7bf832c644debfc06335f6bc6b10e672722761265fe"} Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.636772 4789 scope.go:117] "RemoveContainer" containerID="e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.636877 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7547999-h5qmx" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.655463 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-nrhvq" podStartSLOduration=13.655441992 podStartE2EDuration="13.655441992s" podCreationTimestamp="2025-10-08 14:19:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:16.646936601 +0000 UTC m=+1096.553684103" watchObservedRunningTime="2025-10-08 14:19:16.655441992 +0000 UTC m=+1096.562189484" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.663898 4789 scope.go:117] "RemoveContainer" containerID="783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.676614 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-pr6sr" podStartSLOduration=2.144425949 podStartE2EDuration="10.676596165s" podCreationTimestamp="2025-10-08 14:19:06 +0000 UTC" firstStartedPulling="2025-10-08 14:19:07.71351736 +0000 UTC m=+1087.620264852" lastFinishedPulling="2025-10-08 14:19:16.245687576 +0000 UTC m=+1096.152435068" observedRunningTime="2025-10-08 14:19:16.673437199 +0000 UTC m=+1096.580184681" watchObservedRunningTime="2025-10-08 14:19:16.676596165 +0000 UTC m=+1096.583343657" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.702532 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-dns-svc\") pod \"40054025-fd50-4abc-a3c9-2241ef2be0e9\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.702669 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-config\") pod \"40054025-fd50-4abc-a3c9-2241ef2be0e9\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.703255 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dknt\" (UniqueName: \"kubernetes.io/projected/40054025-fd50-4abc-a3c9-2241ef2be0e9-kube-api-access-7dknt\") pod \"40054025-fd50-4abc-a3c9-2241ef2be0e9\" (UID: \"40054025-fd50-4abc-a3c9-2241ef2be0e9\") " Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.709697 4789 scope.go:117] "RemoveContainer" containerID="e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f" Oct 08 14:19:16 crc kubenswrapper[4789]: E1008 14:19:16.710973 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f\": container with ID starting with e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f not found: ID does not exist" containerID="e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.711100 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f"} err="failed to get container status \"e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f\": rpc error: code = NotFound desc = could not find container \"e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f\": container with ID starting with e695d8827da49e3282543b120c9f503582743da8dfb4a7dcb1b91e45817a596f not found: ID does not exist" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.711127 4789 scope.go:117] "RemoveContainer" containerID="783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b" Oct 08 14:19:16 crc kubenswrapper[4789]: E1008 14:19:16.711724 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b\": container with ID starting with 783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b not found: ID does not exist" containerID="783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.711766 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b"} err="failed to get container status \"783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b\": rpc error: code = NotFound desc = could not find container \"783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b\": container with ID starting with 783fa7e4eb0a9d08c1cd075c417f42c58a8ef91c30ddff81e8ea141760d9108b not found: ID does not exist" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.719206 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40054025-fd50-4abc-a3c9-2241ef2be0e9-kube-api-access-7dknt" (OuterVolumeSpecName: "kube-api-access-7dknt") pod "40054025-fd50-4abc-a3c9-2241ef2be0e9" (UID: "40054025-fd50-4abc-a3c9-2241ef2be0e9"). InnerVolumeSpecName "kube-api-access-7dknt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.806398 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dknt\" (UniqueName: \"kubernetes.io/projected/40054025-fd50-4abc-a3c9-2241ef2be0e9-kube-api-access-7dknt\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.890662 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "40054025-fd50-4abc-a3c9-2241ef2be0e9" (UID: "40054025-fd50-4abc-a3c9-2241ef2be0e9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.910243 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:16 crc kubenswrapper[4789]: I1008 14:19:16.957514 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-config" (OuterVolumeSpecName: "config") pod "40054025-fd50-4abc-a3c9-2241ef2be0e9" (UID: "40054025-fd50-4abc-a3c9-2241ef2be0e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.010774 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40054025-fd50-4abc-a3c9-2241ef2be0e9-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.147920 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.174153 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.272971 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7547999-h5qmx"] Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.283753 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7547999-h5qmx"] Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.315354 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sznk\" (UniqueName: \"kubernetes.io/projected/141f2b14-7eb3-4345-b5ee-ee9666712853-kube-api-access-6sznk\") pod \"141f2b14-7eb3-4345-b5ee-ee9666712853\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.315646 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-ovsdbserver-sb\") pod \"141f2b14-7eb3-4345-b5ee-ee9666712853\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.315755 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-nb\") pod \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.315850 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-config\") pod \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.316010 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-dns-svc\") pod \"141f2b14-7eb3-4345-b5ee-ee9666712853\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.316110 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/a7d4a0e7-d24d-4381-83bc-23b094382b0b-kube-api-access-6jz5f\") pod \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.316363 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-dns-svc\") pod \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.316469 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-sb\") pod \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\" (UID: \"a7d4a0e7-d24d-4381-83bc-23b094382b0b\") " Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.316764 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-config\") pod \"141f2b14-7eb3-4345-b5ee-ee9666712853\" (UID: \"141f2b14-7eb3-4345-b5ee-ee9666712853\") " Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.319937 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d4a0e7-d24d-4381-83bc-23b094382b0b-kube-api-access-6jz5f" (OuterVolumeSpecName: "kube-api-access-6jz5f") pod "a7d4a0e7-d24d-4381-83bc-23b094382b0b" (UID: "a7d4a0e7-d24d-4381-83bc-23b094382b0b"). InnerVolumeSpecName "kube-api-access-6jz5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.319971 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/141f2b14-7eb3-4345-b5ee-ee9666712853-kube-api-access-6sznk" (OuterVolumeSpecName: "kube-api-access-6sznk") pod "141f2b14-7eb3-4345-b5ee-ee9666712853" (UID: "141f2b14-7eb3-4345-b5ee-ee9666712853"). InnerVolumeSpecName "kube-api-access-6sznk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.338506 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a7d4a0e7-d24d-4381-83bc-23b094382b0b" (UID: "a7d4a0e7-d24d-4381-83bc-23b094382b0b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.339099 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7d4a0e7-d24d-4381-83bc-23b094382b0b" (UID: "a7d4a0e7-d24d-4381-83bc-23b094382b0b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.342906 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "141f2b14-7eb3-4345-b5ee-ee9666712853" (UID: "141f2b14-7eb3-4345-b5ee-ee9666712853"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.343541 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-config" (OuterVolumeSpecName: "config") pod "a7d4a0e7-d24d-4381-83bc-23b094382b0b" (UID: "a7d4a0e7-d24d-4381-83bc-23b094382b0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.346694 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-config" (OuterVolumeSpecName: "config") pod "141f2b14-7eb3-4345-b5ee-ee9666712853" (UID: "141f2b14-7eb3-4345-b5ee-ee9666712853"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.354510 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "141f2b14-7eb3-4345-b5ee-ee9666712853" (UID: "141f2b14-7eb3-4345-b5ee-ee9666712853"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.355298 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a7d4a0e7-d24d-4381-83bc-23b094382b0b" (UID: "a7d4a0e7-d24d-4381-83bc-23b094382b0b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.419252 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sznk\" (UniqueName: \"kubernetes.io/projected/141f2b14-7eb3-4345-b5ee-ee9666712853-kube-api-access-6sznk\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.419298 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.419310 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.419322 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.419334 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.419347 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/a7d4a0e7-d24d-4381-83bc-23b094382b0b-kube-api-access-6jz5f\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.419359 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.419370 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7d4a0e7-d24d-4381-83bc-23b094382b0b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.419381 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/141f2b14-7eb3-4345-b5ee-ee9666712853-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.645526 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.645519 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6954cb7f-4mjzq" event={"ID":"a7d4a0e7-d24d-4381-83bc-23b094382b0b","Type":"ContainerDied","Data":"d2b7b1884fe2df817761659d4be64376fc54c12e75c760a6ec591c45c7cecbf1"} Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.646671 4789 scope.go:117] "RemoveContainer" containerID="07f0cdfd9c0b5c8ad03a5ffbe321024b8bbf36542fc5fb9d3f6a416e1859e5e9" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.647756 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" event={"ID":"141f2b14-7eb3-4345-b5ee-ee9666712853","Type":"ContainerDied","Data":"e6a306cb4b1089d254b29a4fbd7c7a910ae1637a913e8a2bb08cd813d40e4d51"} Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.647820 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5677ffc7f-p8lx8" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.650294 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fd23ae36-37d1-458e-ab17-0f0db57e57d3","Type":"ContainerStarted","Data":"31ed1320f3ac2b7c0e43603069d52f334543754a1165850f86ce9ae637eff450"} Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.650649 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.690118 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.045331188 podStartE2EDuration="9.690099253s" podCreationTimestamp="2025-10-08 14:19:08 +0000 UTC" firstStartedPulling="2025-10-08 14:19:09.598416764 +0000 UTC m=+1089.505164256" lastFinishedPulling="2025-10-08 14:19:16.243184829 +0000 UTC m=+1096.149932321" observedRunningTime="2025-10-08 14:19:17.674700026 +0000 UTC m=+1097.581447538" watchObservedRunningTime="2025-10-08 14:19:17.690099253 +0000 UTC m=+1097.596846745" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.696138 4789 scope.go:117] "RemoveContainer" containerID="88a1e2db7673690b02ff31d512303bbee22976fb2635059c8cb0a0c558f1ac89" Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.793075 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f6954cb7f-4mjzq"] Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.799068 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f6954cb7f-4mjzq"] Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.826583 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5677ffc7f-p8lx8"] Oct 08 14:19:17 crc kubenswrapper[4789]: I1008 14:19:17.834515 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5677ffc7f-p8lx8"] Oct 08 14:19:18 crc kubenswrapper[4789]: I1008 14:19:18.739666 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="141f2b14-7eb3-4345-b5ee-ee9666712853" path="/var/lib/kubelet/pods/141f2b14-7eb3-4345-b5ee-ee9666712853/volumes" Oct 08 14:19:18 crc kubenswrapper[4789]: I1008 14:19:18.740382 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40054025-fd50-4abc-a3c9-2241ef2be0e9" path="/var/lib/kubelet/pods/40054025-fd50-4abc-a3c9-2241ef2be0e9/volumes" Oct 08 14:19:18 crc kubenswrapper[4789]: I1008 14:19:18.741175 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7d4a0e7-d24d-4381-83bc-23b094382b0b" path="/var/lib/kubelet/pods/a7d4a0e7-d24d-4381-83bc-23b094382b0b/volumes" Oct 08 14:19:19 crc kubenswrapper[4789]: I1008 14:19:19.873765 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 08 14:19:19 crc kubenswrapper[4789]: I1008 14:19:19.934357 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 08 14:19:21 crc kubenswrapper[4789]: I1008 14:19:21.705218 4789 generic.go:334] "Generic (PLEG): container finished" podID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerID="e3f4a50d2a44c2277c7f6e583c4eeef4fadd0de7449e696ac9b577620783eb3f" exitCode=0 Oct 08 14:19:21 crc kubenswrapper[4789]: I1008 14:19:21.705254 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c9336a8d-1117-4da6-a21b-4070f722adc9","Type":"ContainerDied","Data":"e3f4a50d2a44c2277c7f6e583c4eeef4fadd0de7449e696ac9b577620783eb3f"} Oct 08 14:19:21 crc kubenswrapper[4789]: I1008 14:19:21.708493 4789 generic.go:334] "Generic (PLEG): container finished" podID="55fc0a2c-48c9-4508-ae9d-5350f3298621" containerID="6a761c95276f91455b1d8de776b172e2014ff8e5c08e5130aa1c771e2d35c9ab" exitCode=0 Oct 08 14:19:21 crc kubenswrapper[4789]: I1008 14:19:21.708552 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"55fc0a2c-48c9-4508-ae9d-5350f3298621","Type":"ContainerDied","Data":"6a761c95276f91455b1d8de776b172e2014ff8e5c08e5130aa1c771e2d35c9ab"} Oct 08 14:19:21 crc kubenswrapper[4789]: I1008 14:19:21.712151 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerStarted","Data":"bdd9e59a2c8f14973ba2ce56ccc6945b2ee66964a0e7da6de0f0e807a11f5d36"} Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.141701 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.222155 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.230225 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:22 crc kubenswrapper[4789]: E1008 14:19:22.230432 4789 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 14:19:22 crc kubenswrapper[4789]: E1008 14:19:22.230453 4789 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 14:19:22 crc kubenswrapper[4789]: E1008 14:19:22.230506 4789 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift podName:0462c12d-8df3-4734-9c2e-a925c9c06b9e nodeName:}" failed. No retries permitted until 2025-10-08 14:19:38.230491836 +0000 UTC m=+1118.137239328 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift") pod "swift-storage-0" (UID: "0462c12d-8df3-4734-9c2e-a925c9c06b9e") : configmap "swift-ring-files" not found Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.569129 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-kxpvf"] Oct 08 14:19:22 crc kubenswrapper[4789]: E1008 14:19:22.569528 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="141f2b14-7eb3-4345-b5ee-ee9666712853" containerName="init" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.569550 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="141f2b14-7eb3-4345-b5ee-ee9666712853" containerName="init" Oct 08 14:19:22 crc kubenswrapper[4789]: E1008 14:19:22.569575 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d4a0e7-d24d-4381-83bc-23b094382b0b" containerName="init" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.569602 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d4a0e7-d24d-4381-83bc-23b094382b0b" containerName="init" Oct 08 14:19:22 crc kubenswrapper[4789]: E1008 14:19:22.569619 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40054025-fd50-4abc-a3c9-2241ef2be0e9" containerName="init" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.569627 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="40054025-fd50-4abc-a3c9-2241ef2be0e9" containerName="init" Oct 08 14:19:22 crc kubenswrapper[4789]: E1008 14:19:22.569637 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40054025-fd50-4abc-a3c9-2241ef2be0e9" containerName="dnsmasq-dns" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.569644 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="40054025-fd50-4abc-a3c9-2241ef2be0e9" containerName="dnsmasq-dns" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.569858 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="40054025-fd50-4abc-a3c9-2241ef2be0e9" containerName="dnsmasq-dns" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.569879 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d4a0e7-d24d-4381-83bc-23b094382b0b" containerName="init" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.569898 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="141f2b14-7eb3-4345-b5ee-ee9666712853" containerName="init" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.570619 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kxpvf" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.574183 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-kxpvf"] Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.721288 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c9336a8d-1117-4da6-a21b-4070f722adc9","Type":"ContainerStarted","Data":"2d145f358f5e7b194fe1775e804a9988dab919e072ce986e70c64440037eba38"} Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.721809 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.723477 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"55fc0a2c-48c9-4508-ae9d-5350f3298621","Type":"ContainerStarted","Data":"785d1fc49a55a8d566c8754a267db4182e5556932925e12f766f9b1294a9b3dc"} Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.739017 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg58t\" (UniqueName: \"kubernetes.io/projected/b8f44a9a-deb5-478b-8414-f103280bdf3f-kube-api-access-bg58t\") pod \"placement-db-create-kxpvf\" (UID: \"b8f44a9a-deb5-478b-8414-f103280bdf3f\") " pod="openstack/placement-db-create-kxpvf" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.748931 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.60118574 podStartE2EDuration="55.748912436s" podCreationTimestamp="2025-10-08 14:18:27 +0000 UTC" firstStartedPulling="2025-10-08 14:18:29.268358605 +0000 UTC m=+1049.175106097" lastFinishedPulling="2025-10-08 14:18:47.416085301 +0000 UTC m=+1067.322832793" observedRunningTime="2025-10-08 14:19:22.742546174 +0000 UTC m=+1102.649293666" watchObservedRunningTime="2025-10-08 14:19:22.748912436 +0000 UTC m=+1102.655659928" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.782289 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-notifications-server-0" podStartSLOduration=45.592330045 podStartE2EDuration="55.78227209s" podCreationTimestamp="2025-10-08 14:18:27 +0000 UTC" firstStartedPulling="2025-10-08 14:18:37.20488645 +0000 UTC m=+1057.111633932" lastFinishedPulling="2025-10-08 14:18:47.394828485 +0000 UTC m=+1067.301575977" observedRunningTime="2025-10-08 14:19:22.776072772 +0000 UTC m=+1102.682820264" watchObservedRunningTime="2025-10-08 14:19:22.78227209 +0000 UTC m=+1102.689019582" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.840858 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg58t\" (UniqueName: \"kubernetes.io/projected/b8f44a9a-deb5-478b-8414-f103280bdf3f-kube-api-access-bg58t\") pod \"placement-db-create-kxpvf\" (UID: \"b8f44a9a-deb5-478b-8414-f103280bdf3f\") " pod="openstack/placement-db-create-kxpvf" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.877439 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg58t\" (UniqueName: \"kubernetes.io/projected/b8f44a9a-deb5-478b-8414-f103280bdf3f-kube-api-access-bg58t\") pod \"placement-db-create-kxpvf\" (UID: \"b8f44a9a-deb5-478b-8414-f103280bdf3f\") " pod="openstack/placement-db-create-kxpvf" Oct 08 14:19:22 crc kubenswrapper[4789]: I1008 14:19:22.902808 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kxpvf" Oct 08 14:19:23 crc kubenswrapper[4789]: I1008 14:19:23.436056 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-kxpvf"] Oct 08 14:19:23 crc kubenswrapper[4789]: W1008 14:19:23.478566 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8f44a9a_deb5_478b_8414_f103280bdf3f.slice/crio-6719f5a0d9ca0dbd0a9c6d115690ad3dd45c9739ab2625294782e6bf6d4b6e03 WatchSource:0}: Error finding container 6719f5a0d9ca0dbd0a9c6d115690ad3dd45c9739ab2625294782e6bf6d4b6e03: Status 404 returned error can't find the container with id 6719f5a0d9ca0dbd0a9c6d115690ad3dd45c9739ab2625294782e6bf6d4b6e03 Oct 08 14:19:23 crc kubenswrapper[4789]: I1008 14:19:23.733362 4789 generic.go:334] "Generic (PLEG): container finished" podID="05192759-43d6-4ef0-b6f0-6284e2435317" containerID="257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c" exitCode=0 Oct 08 14:19:23 crc kubenswrapper[4789]: I1008 14:19:23.733439 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"05192759-43d6-4ef0-b6f0-6284e2435317","Type":"ContainerDied","Data":"257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c"} Oct 08 14:19:23 crc kubenswrapper[4789]: I1008 14:19:23.737845 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kxpvf" event={"ID":"b8f44a9a-deb5-478b-8414-f103280bdf3f","Type":"ContainerStarted","Data":"62ded9631a4b6121d80f881faf6895ef2d0c3dd4da082bd8cd584d3828f5d23d"} Oct 08 14:19:23 crc kubenswrapper[4789]: I1008 14:19:23.737903 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kxpvf" event={"ID":"b8f44a9a-deb5-478b-8414-f103280bdf3f","Type":"ContainerStarted","Data":"6719f5a0d9ca0dbd0a9c6d115690ad3dd45c9739ab2625294782e6bf6d4b6e03"} Oct 08 14:19:23 crc kubenswrapper[4789]: I1008 14:19:23.780872 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-kxpvf" podStartSLOduration=1.7808503629999999 podStartE2EDuration="1.780850363s" podCreationTimestamp="2025-10-08 14:19:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:23.772701232 +0000 UTC m=+1103.679448724" watchObservedRunningTime="2025-10-08 14:19:23.780850363 +0000 UTC m=+1103.687597865" Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.745539 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerStarted","Data":"134be9867b022fdfd415b2226c4968610d9f9cc7451cb4ad293311c57bc73b38"} Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.747761 4789 generic.go:334] "Generic (PLEG): container finished" podID="b8f44a9a-deb5-478b-8414-f103280bdf3f" containerID="62ded9631a4b6121d80f881faf6895ef2d0c3dd4da082bd8cd584d3828f5d23d" exitCode=0 Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.747811 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kxpvf" event={"ID":"b8f44a9a-deb5-478b-8414-f103280bdf3f","Type":"ContainerDied","Data":"62ded9631a4b6121d80f881faf6895ef2d0c3dd4da082bd8cd584d3828f5d23d"} Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.749533 4789 generic.go:334] "Generic (PLEG): container finished" podID="a7b17c27-5518-480a-875f-091ae1dd3606" containerID="ab0b023a33b800df6b2d7de7569d43ca0fa1aef5e07dee2962ca4acbea9445eb" exitCode=0 Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.749607 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pr6sr" event={"ID":"a7b17c27-5518-480a-875f-091ae1dd3606","Type":"ContainerDied","Data":"ab0b023a33b800df6b2d7de7569d43ca0fa1aef5e07dee2962ca4acbea9445eb"} Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.751766 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"05192759-43d6-4ef0-b6f0-6284e2435317","Type":"ContainerStarted","Data":"bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b"} Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.752048 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.815109 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=56.815090373 podStartE2EDuration="56.815090373s" podCreationTimestamp="2025-10-08 14:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:24.814338723 +0000 UTC m=+1104.721086235" watchObservedRunningTime="2025-10-08 14:19:24.815090373 +0000 UTC m=+1104.721837875" Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.961597 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-f6hnn"] Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.963081 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-f6hnn" Oct 08 14:19:24 crc kubenswrapper[4789]: I1008 14:19:24.975210 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-f6hnn"] Oct 08 14:19:25 crc kubenswrapper[4789]: I1008 14:19:25.108934 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c72sb\" (UniqueName: \"kubernetes.io/projected/14b540ee-ce54-430d-8313-d7b33bcdb1a7-kube-api-access-c72sb\") pod \"watcher-db-create-f6hnn\" (UID: \"14b540ee-ce54-430d-8313-d7b33bcdb1a7\") " pod="openstack/watcher-db-create-f6hnn" Oct 08 14:19:25 crc kubenswrapper[4789]: I1008 14:19:25.210410 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c72sb\" (UniqueName: \"kubernetes.io/projected/14b540ee-ce54-430d-8313-d7b33bcdb1a7-kube-api-access-c72sb\") pod \"watcher-db-create-f6hnn\" (UID: \"14b540ee-ce54-430d-8313-d7b33bcdb1a7\") " pod="openstack/watcher-db-create-f6hnn" Oct 08 14:19:25 crc kubenswrapper[4789]: I1008 14:19:25.241066 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c72sb\" (UniqueName: \"kubernetes.io/projected/14b540ee-ce54-430d-8313-d7b33bcdb1a7-kube-api-access-c72sb\") pod \"watcher-db-create-f6hnn\" (UID: \"14b540ee-ce54-430d-8313-d7b33bcdb1a7\") " pod="openstack/watcher-db-create-f6hnn" Oct 08 14:19:25 crc kubenswrapper[4789]: I1008 14:19:25.317517 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-f6hnn" Oct 08 14:19:25 crc kubenswrapper[4789]: I1008 14:19:25.757956 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-f6hnn"] Oct 08 14:19:25 crc kubenswrapper[4789]: W1008 14:19:25.761399 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14b540ee_ce54_430d_8313_d7b33bcdb1a7.slice/crio-bb4d221725583f272b42d89829c79bbfecaaaf4141b7aa6a79606b9998b12d70 WatchSource:0}: Error finding container bb4d221725583f272b42d89829c79bbfecaaaf4141b7aa6a79606b9998b12d70: Status 404 returned error can't find the container with id bb4d221725583f272b42d89829c79bbfecaaaf4141b7aa6a79606b9998b12d70 Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.084594 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kxpvf" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.155598 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.226319 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-swiftconf\") pod \"a7b17c27-5518-480a-875f-091ae1dd3606\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.227169 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgv4r\" (UniqueName: \"kubernetes.io/projected/a7b17c27-5518-480a-875f-091ae1dd3606-kube-api-access-qgv4r\") pod \"a7b17c27-5518-480a-875f-091ae1dd3606\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.227211 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-dispersionconf\") pod \"a7b17c27-5518-480a-875f-091ae1dd3606\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.227705 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg58t\" (UniqueName: \"kubernetes.io/projected/b8f44a9a-deb5-478b-8414-f103280bdf3f-kube-api-access-bg58t\") pod \"b8f44a9a-deb5-478b-8414-f103280bdf3f\" (UID: \"b8f44a9a-deb5-478b-8414-f103280bdf3f\") " Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.227734 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-scripts\") pod \"a7b17c27-5518-480a-875f-091ae1dd3606\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.227762 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-combined-ca-bundle\") pod \"a7b17c27-5518-480a-875f-091ae1dd3606\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.227824 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7b17c27-5518-480a-875f-091ae1dd3606-etc-swift\") pod \"a7b17c27-5518-480a-875f-091ae1dd3606\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.227863 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-ring-data-devices\") pod \"a7b17c27-5518-480a-875f-091ae1dd3606\" (UID: \"a7b17c27-5518-480a-875f-091ae1dd3606\") " Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.228669 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a7b17c27-5518-480a-875f-091ae1dd3606" (UID: "a7b17c27-5518-480a-875f-091ae1dd3606"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.229580 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7b17c27-5518-480a-875f-091ae1dd3606-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a7b17c27-5518-480a-875f-091ae1dd3606" (UID: "a7b17c27-5518-480a-875f-091ae1dd3606"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.231889 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7b17c27-5518-480a-875f-091ae1dd3606-kube-api-access-qgv4r" (OuterVolumeSpecName: "kube-api-access-qgv4r") pod "a7b17c27-5518-480a-875f-091ae1dd3606" (UID: "a7b17c27-5518-480a-875f-091ae1dd3606"). InnerVolumeSpecName "kube-api-access-qgv4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.232724 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8f44a9a-deb5-478b-8414-f103280bdf3f-kube-api-access-bg58t" (OuterVolumeSpecName: "kube-api-access-bg58t") pod "b8f44a9a-deb5-478b-8414-f103280bdf3f" (UID: "b8f44a9a-deb5-478b-8414-f103280bdf3f"). InnerVolumeSpecName "kube-api-access-bg58t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.249617 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-scripts" (OuterVolumeSpecName: "scripts") pod "a7b17c27-5518-480a-875f-091ae1dd3606" (UID: "a7b17c27-5518-480a-875f-091ae1dd3606"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.249722 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a7b17c27-5518-480a-875f-091ae1dd3606" (UID: "a7b17c27-5518-480a-875f-091ae1dd3606"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.255670 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a7b17c27-5518-480a-875f-091ae1dd3606" (UID: "a7b17c27-5518-480a-875f-091ae1dd3606"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.259813 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7b17c27-5518-480a-875f-091ae1dd3606" (UID: "a7b17c27-5518-480a-875f-091ae1dd3606"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.330209 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgv4r\" (UniqueName: \"kubernetes.io/projected/a7b17c27-5518-480a-875f-091ae1dd3606-kube-api-access-qgv4r\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.330257 4789 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.330270 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg58t\" (UniqueName: \"kubernetes.io/projected/b8f44a9a-deb5-478b-8414-f103280bdf3f-kube-api-access-bg58t\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.330283 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.330294 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.330307 4789 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a7b17c27-5518-480a-875f-091ae1dd3606-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.330317 4789 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a7b17c27-5518-480a-875f-091ae1dd3606-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.330329 4789 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a7b17c27-5518-480a-875f-091ae1dd3606-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.432686 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.432738 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.772027 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pr6sr" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.773562 4789 generic.go:334] "Generic (PLEG): container finished" podID="14b540ee-ce54-430d-8313-d7b33bcdb1a7" containerID="4988a5e1718962ab0e50dc3e6bd1fd6823f07a8bd11141afebb5f985202e9af4" exitCode=0 Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.786607 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pr6sr" event={"ID":"a7b17c27-5518-480a-875f-091ae1dd3606","Type":"ContainerDied","Data":"36fee5c2a6ddff8971e8cd8dd3581e8ae6d06ec6b619ad2528689fe312254fb5"} Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.786695 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36fee5c2a6ddff8971e8cd8dd3581e8ae6d06ec6b619ad2528689fe312254fb5" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.786716 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-f6hnn" event={"ID":"14b540ee-ce54-430d-8313-d7b33bcdb1a7","Type":"ContainerDied","Data":"4988a5e1718962ab0e50dc3e6bd1fd6823f07a8bd11141afebb5f985202e9af4"} Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.786734 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-f6hnn" event={"ID":"14b540ee-ce54-430d-8313-d7b33bcdb1a7","Type":"ContainerStarted","Data":"bb4d221725583f272b42d89829c79bbfecaaaf4141b7aa6a79606b9998b12d70"} Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.796533 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kxpvf" event={"ID":"b8f44a9a-deb5-478b-8414-f103280bdf3f","Type":"ContainerDied","Data":"6719f5a0d9ca0dbd0a9c6d115690ad3dd45c9739ab2625294782e6bf6d4b6e03"} Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.796583 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6719f5a0d9ca0dbd0a9c6d115690ad3dd45c9739ab2625294782e6bf6d4b6e03" Oct 08 14:19:26 crc kubenswrapper[4789]: I1008 14:19:26.796732 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kxpvf" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.170909 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-f6hnn" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.261855 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c72sb\" (UniqueName: \"kubernetes.io/projected/14b540ee-ce54-430d-8313-d7b33bcdb1a7-kube-api-access-c72sb\") pod \"14b540ee-ce54-430d-8313-d7b33bcdb1a7\" (UID: \"14b540ee-ce54-430d-8313-d7b33bcdb1a7\") " Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.276322 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zrdhb" podUID="91ddb1cc-17c9-4f38-ac8a-d78d10798178" containerName="ovn-controller" probeResult="failure" output=< Oct 08 14:19:28 crc kubenswrapper[4789]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 14:19:28 crc kubenswrapper[4789]: > Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.285245 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14b540ee-ce54-430d-8313-d7b33bcdb1a7-kube-api-access-c72sb" (OuterVolumeSpecName: "kube-api-access-c72sb") pod "14b540ee-ce54-430d-8313-d7b33bcdb1a7" (UID: "14b540ee-ce54-430d-8313-d7b33bcdb1a7"). InnerVolumeSpecName "kube-api-access-c72sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.303614 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.307875 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-tkv8v" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.364032 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c72sb\" (UniqueName: \"kubernetes.io/projected/14b540ee-ce54-430d-8313-d7b33bcdb1a7-kube-api-access-c72sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.527508 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zrdhb-config-9cfb4"] Oct 08 14:19:28 crc kubenswrapper[4789]: E1008 14:19:28.527919 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14b540ee-ce54-430d-8313-d7b33bcdb1a7" containerName="mariadb-database-create" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.527940 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="14b540ee-ce54-430d-8313-d7b33bcdb1a7" containerName="mariadb-database-create" Oct 08 14:19:28 crc kubenswrapper[4789]: E1008 14:19:28.527965 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b17c27-5518-480a-875f-091ae1dd3606" containerName="swift-ring-rebalance" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.527975 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b17c27-5518-480a-875f-091ae1dd3606" containerName="swift-ring-rebalance" Oct 08 14:19:28 crc kubenswrapper[4789]: E1008 14:19:28.528003 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8f44a9a-deb5-478b-8414-f103280bdf3f" containerName="mariadb-database-create" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.528012 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8f44a9a-deb5-478b-8414-f103280bdf3f" containerName="mariadb-database-create" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.528203 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8f44a9a-deb5-478b-8414-f103280bdf3f" containerName="mariadb-database-create" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.528232 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="14b540ee-ce54-430d-8313-d7b33bcdb1a7" containerName="mariadb-database-create" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.528241 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b17c27-5518-480a-875f-091ae1dd3606" containerName="swift-ring-rebalance" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.528927 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.532760 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.545108 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrdhb-config-9cfb4"] Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.567493 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.567556 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-additional-scripts\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.567644 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shkxh\" (UniqueName: \"kubernetes.io/projected/c47b01d8-8a47-43fc-9780-242aa88984da-kube-api-access-shkxh\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.567700 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run-ovn\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.567735 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-log-ovn\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.567761 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-scripts\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669046 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-additional-scripts\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669168 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shkxh\" (UniqueName: \"kubernetes.io/projected/c47b01d8-8a47-43fc-9780-242aa88984da-kube-api-access-shkxh\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669228 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run-ovn\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669259 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-log-ovn\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669284 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-scripts\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669355 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669577 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669576 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-log-ovn\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669585 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run-ovn\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.669780 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-additional-scripts\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.671484 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-scripts\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.686687 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shkxh\" (UniqueName: \"kubernetes.io/projected/c47b01d8-8a47-43fc-9780-242aa88984da-kube-api-access-shkxh\") pod \"ovn-controller-zrdhb-config-9cfb4\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.814049 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerStarted","Data":"0862d5c6b527f4d758ce54d7e6f4c60e4a85b8e64e1ccfdd9b63f944c7ef981c"} Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.815856 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-f6hnn" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.815930 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-f6hnn" event={"ID":"14b540ee-ce54-430d-8313-d7b33bcdb1a7","Type":"ContainerDied","Data":"bb4d221725583f272b42d89829c79bbfecaaaf4141b7aa6a79606b9998b12d70"} Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.815955 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb4d221725583f272b42d89829c79bbfecaaaf4141b7aa6a79606b9998b12d70" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.839192 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.015567362 podStartE2EDuration="54.839169673s" podCreationTimestamp="2025-10-08 14:18:34 +0000 UTC" firstStartedPulling="2025-10-08 14:18:49.008627112 +0000 UTC m=+1068.915374604" lastFinishedPulling="2025-10-08 14:19:27.832229423 +0000 UTC m=+1107.738976915" observedRunningTime="2025-10-08 14:19:28.835538955 +0000 UTC m=+1108.742286447" watchObservedRunningTime="2025-10-08 14:19:28.839169673 +0000 UTC m=+1108.745917175" Oct 08 14:19:28 crc kubenswrapper[4789]: I1008 14:19:28.847322 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:29 crc kubenswrapper[4789]: I1008 14:19:29.150528 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:19:29 crc kubenswrapper[4789]: I1008 14:19:29.177439 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 08 14:19:29 crc kubenswrapper[4789]: I1008 14:19:29.371411 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zrdhb-config-9cfb4"] Oct 08 14:19:29 crc kubenswrapper[4789]: W1008 14:19:29.375154 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc47b01d8_8a47_43fc_9780_242aa88984da.slice/crio-70dc57b08241bf755c5e6c561e7e918e49394e263686ad0c1befa7b182700bcd WatchSource:0}: Error finding container 70dc57b08241bf755c5e6c561e7e918e49394e263686ad0c1befa7b182700bcd: Status 404 returned error can't find the container with id 70dc57b08241bf755c5e6c561e7e918e49394e263686ad0c1befa7b182700bcd Oct 08 14:19:29 crc kubenswrapper[4789]: I1008 14:19:29.823289 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrdhb-config-9cfb4" event={"ID":"c47b01d8-8a47-43fc-9780-242aa88984da","Type":"ContainerStarted","Data":"db1238911101afd5fd59caa03415ea3bae71b614c0004c9b526ebc31465a6089"} Oct 08 14:19:29 crc kubenswrapper[4789]: I1008 14:19:29.823627 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrdhb-config-9cfb4" event={"ID":"c47b01d8-8a47-43fc-9780-242aa88984da","Type":"ContainerStarted","Data":"70dc57b08241bf755c5e6c561e7e918e49394e263686ad0c1befa7b182700bcd"} Oct 08 14:19:29 crc kubenswrapper[4789]: I1008 14:19:29.844157 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zrdhb-config-9cfb4" podStartSLOduration=1.84413816 podStartE2EDuration="1.84413816s" podCreationTimestamp="2025-10-08 14:19:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:29.843002189 +0000 UTC m=+1109.749749681" watchObservedRunningTime="2025-10-08 14:19:29.84413816 +0000 UTC m=+1109.750885652" Oct 08 14:19:30 crc kubenswrapper[4789]: I1008 14:19:30.833033 4789 generic.go:334] "Generic (PLEG): container finished" podID="c47b01d8-8a47-43fc-9780-242aa88984da" containerID="db1238911101afd5fd59caa03415ea3bae71b614c0004c9b526ebc31465a6089" exitCode=0 Oct 08 14:19:30 crc kubenswrapper[4789]: I1008 14:19:30.833085 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrdhb-config-9cfb4" event={"ID":"c47b01d8-8a47-43fc-9780-242aa88984da","Type":"ContainerDied","Data":"db1238911101afd5fd59caa03415ea3bae71b614c0004c9b526ebc31465a6089"} Oct 08 14:19:31 crc kubenswrapper[4789]: I1008 14:19:31.129438 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.143616 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-pd9bc"] Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.145165 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pd9bc" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.169499 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-pd9bc"] Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.193603 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.237366 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m79c\" (UniqueName: \"kubernetes.io/projected/bdab5d2d-97d5-4c68-ad3e-2589adfcadd4-kube-api-access-2m79c\") pod \"keystone-db-create-pd9bc\" (UID: \"bdab5d2d-97d5-4c68-ad3e-2589adfcadd4\") " pod="openstack/keystone-db-create-pd9bc" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.338227 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run\") pod \"c47b01d8-8a47-43fc-9780-242aa88984da\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.338388 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run-ovn\") pod \"c47b01d8-8a47-43fc-9780-242aa88984da\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.338385 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run" (OuterVolumeSpecName: "var-run") pod "c47b01d8-8a47-43fc-9780-242aa88984da" (UID: "c47b01d8-8a47-43fc-9780-242aa88984da"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.338483 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c47b01d8-8a47-43fc-9780-242aa88984da" (UID: "c47b01d8-8a47-43fc-9780-242aa88984da"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.338542 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-additional-scripts\") pod \"c47b01d8-8a47-43fc-9780-242aa88984da\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.339340 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c47b01d8-8a47-43fc-9780-242aa88984da" (UID: "c47b01d8-8a47-43fc-9780-242aa88984da"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.339515 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-log-ovn\") pod \"c47b01d8-8a47-43fc-9780-242aa88984da\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.339583 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c47b01d8-8a47-43fc-9780-242aa88984da" (UID: "c47b01d8-8a47-43fc-9780-242aa88984da"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.339634 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shkxh\" (UniqueName: \"kubernetes.io/projected/c47b01d8-8a47-43fc-9780-242aa88984da-kube-api-access-shkxh\") pod \"c47b01d8-8a47-43fc-9780-242aa88984da\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.340365 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-scripts\") pod \"c47b01d8-8a47-43fc-9780-242aa88984da\" (UID: \"c47b01d8-8a47-43fc-9780-242aa88984da\") " Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.340788 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m79c\" (UniqueName: \"kubernetes.io/projected/bdab5d2d-97d5-4c68-ad3e-2589adfcadd4-kube-api-access-2m79c\") pod \"keystone-db-create-pd9bc\" (UID: \"bdab5d2d-97d5-4c68-ad3e-2589adfcadd4\") " pod="openstack/keystone-db-create-pd9bc" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.341006 4789 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.341029 4789 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.341043 4789 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c47b01d8-8a47-43fc-9780-242aa88984da-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.341056 4789 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.341153 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-scripts" (OuterVolumeSpecName: "scripts") pod "c47b01d8-8a47-43fc-9780-242aa88984da" (UID: "c47b01d8-8a47-43fc-9780-242aa88984da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.345544 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c47b01d8-8a47-43fc-9780-242aa88984da-kube-api-access-shkxh" (OuterVolumeSpecName: "kube-api-access-shkxh") pod "c47b01d8-8a47-43fc-9780-242aa88984da" (UID: "c47b01d8-8a47-43fc-9780-242aa88984da"). InnerVolumeSpecName "kube-api-access-shkxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.359455 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m79c\" (UniqueName: \"kubernetes.io/projected/bdab5d2d-97d5-4c68-ad3e-2589adfcadd4-kube-api-access-2m79c\") pod \"keystone-db-create-pd9bc\" (UID: \"bdab5d2d-97d5-4c68-ad3e-2589adfcadd4\") " pod="openstack/keystone-db-create-pd9bc" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.442675 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shkxh\" (UniqueName: \"kubernetes.io/projected/c47b01d8-8a47-43fc-9780-242aa88984da-kube-api-access-shkxh\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.442714 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c47b01d8-8a47-43fc-9780-242aa88984da-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.506343 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pd9bc" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.581839 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-2d7c-account-create-cgqlh"] Oct 08 14:19:32 crc kubenswrapper[4789]: E1008 14:19:32.582504 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c47b01d8-8a47-43fc-9780-242aa88984da" containerName="ovn-config" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.582522 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c47b01d8-8a47-43fc-9780-242aa88984da" containerName="ovn-config" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.582713 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c47b01d8-8a47-43fc-9780-242aa88984da" containerName="ovn-config" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.583271 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d7c-account-create-cgqlh" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.585206 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.604016 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2d7c-account-create-cgqlh"] Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.646451 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv6wr\" (UniqueName: \"kubernetes.io/projected/dbd72997-3d15-40ee-adfc-b1ffdf7667c9-kube-api-access-jv6wr\") pod \"placement-2d7c-account-create-cgqlh\" (UID: \"dbd72997-3d15-40ee-adfc-b1ffdf7667c9\") " pod="openstack/placement-2d7c-account-create-cgqlh" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.748183 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv6wr\" (UniqueName: \"kubernetes.io/projected/dbd72997-3d15-40ee-adfc-b1ffdf7667c9-kube-api-access-jv6wr\") pod \"placement-2d7c-account-create-cgqlh\" (UID: \"dbd72997-3d15-40ee-adfc-b1ffdf7667c9\") " pod="openstack/placement-2d7c-account-create-cgqlh" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.768881 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv6wr\" (UniqueName: \"kubernetes.io/projected/dbd72997-3d15-40ee-adfc-b1ffdf7667c9-kube-api-access-jv6wr\") pod \"placement-2d7c-account-create-cgqlh\" (UID: \"dbd72997-3d15-40ee-adfc-b1ffdf7667c9\") " pod="openstack/placement-2d7c-account-create-cgqlh" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.852593 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zrdhb-config-9cfb4" event={"ID":"c47b01d8-8a47-43fc-9780-242aa88984da","Type":"ContainerDied","Data":"70dc57b08241bf755c5e6c561e7e918e49394e263686ad0c1befa7b182700bcd"} Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.852635 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70dc57b08241bf755c5e6c561e7e918e49394e263686ad0c1befa7b182700bcd" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.852668 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zrdhb-config-9cfb4" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.907426 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d7c-account-create-cgqlh" Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.946828 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zrdhb-config-9cfb4"] Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.957386 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zrdhb-config-9cfb4"] Oct 08 14:19:32 crc kubenswrapper[4789]: I1008 14:19:32.978084 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-pd9bc"] Oct 08 14:19:32 crc kubenswrapper[4789]: W1008 14:19:32.990553 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdab5d2d_97d5_4c68_ad3e_2589adfcadd4.slice/crio-d0682ae32a29e35bafc73aa23f74ee29be604f5cd1f1807bd91a0c1752bb9503 WatchSource:0}: Error finding container d0682ae32a29e35bafc73aa23f74ee29be604f5cd1f1807bd91a0c1752bb9503: Status 404 returned error can't find the container with id d0682ae32a29e35bafc73aa23f74ee29be604f5cd1f1807bd91a0c1752bb9503 Oct 08 14:19:33 crc kubenswrapper[4789]: I1008 14:19:33.265619 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-zrdhb" Oct 08 14:19:33 crc kubenswrapper[4789]: I1008 14:19:33.463264 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2d7c-account-create-cgqlh"] Oct 08 14:19:33 crc kubenswrapper[4789]: I1008 14:19:33.861430 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d7c-account-create-cgqlh" event={"ID":"dbd72997-3d15-40ee-adfc-b1ffdf7667c9","Type":"ContainerStarted","Data":"91d94b4b082eab6d7ad0744082975e0ad80411542953186d636080a1b2e89707"} Oct 08 14:19:33 crc kubenswrapper[4789]: I1008 14:19:33.862334 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pd9bc" event={"ID":"bdab5d2d-97d5-4c68-ad3e-2589adfcadd4","Type":"ContainerStarted","Data":"d0682ae32a29e35bafc73aa23f74ee29be604f5cd1f1807bd91a0c1752bb9503"} Oct 08 14:19:34 crc kubenswrapper[4789]: I1008 14:19:34.741218 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c47b01d8-8a47-43fc-9780-242aa88984da" path="/var/lib/kubelet/pods/c47b01d8-8a47-43fc-9780-242aa88984da/volumes" Oct 08 14:19:34 crc kubenswrapper[4789]: I1008 14:19:34.871820 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pd9bc" event={"ID":"bdab5d2d-97d5-4c68-ad3e-2589adfcadd4","Type":"ContainerStarted","Data":"e2a9d30875924c891b81d7e19dd1b2f459d6c0a4e8b702e049efaf70f4cc4f38"} Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.087205 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-7de8-account-create-4zc9n"] Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.088860 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-7de8-account-create-4zc9n" Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.090939 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.102256 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-7de8-account-create-4zc9n"] Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.189758 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcsvp\" (UniqueName: \"kubernetes.io/projected/d3f0c4cf-3ebe-440d-a47a-ad8966d5087e-kube-api-access-dcsvp\") pod \"watcher-7de8-account-create-4zc9n\" (UID: \"d3f0c4cf-3ebe-440d-a47a-ad8966d5087e\") " pod="openstack/watcher-7de8-account-create-4zc9n" Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.291967 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcsvp\" (UniqueName: \"kubernetes.io/projected/d3f0c4cf-3ebe-440d-a47a-ad8966d5087e-kube-api-access-dcsvp\") pod \"watcher-7de8-account-create-4zc9n\" (UID: \"d3f0c4cf-3ebe-440d-a47a-ad8966d5087e\") " pod="openstack/watcher-7de8-account-create-4zc9n" Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.314607 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcsvp\" (UniqueName: \"kubernetes.io/projected/d3f0c4cf-3ebe-440d-a47a-ad8966d5087e-kube-api-access-dcsvp\") pod \"watcher-7de8-account-create-4zc9n\" (UID: \"d3f0c4cf-3ebe-440d-a47a-ad8966d5087e\") " pod="openstack/watcher-7de8-account-create-4zc9n" Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.412032 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-7de8-account-create-4zc9n" Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.841314 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-7de8-account-create-4zc9n"] Oct 08 14:19:35 crc kubenswrapper[4789]: I1008 14:19:35.882024 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-7de8-account-create-4zc9n" event={"ID":"d3f0c4cf-3ebe-440d-a47a-ad8966d5087e","Type":"ContainerStarted","Data":"f56ec899cacdd3a017d2dd4d6cee4ccfcf99f8e1fb40ab6d642f97ce307aa2b5"} Oct 08 14:19:36 crc kubenswrapper[4789]: I1008 14:19:36.130335 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:36 crc kubenswrapper[4789]: I1008 14:19:36.133460 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:36 crc kubenswrapper[4789]: I1008 14:19:36.892815 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:37 crc kubenswrapper[4789]: I1008 14:19:37.901812 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-7de8-account-create-4zc9n" event={"ID":"d3f0c4cf-3ebe-440d-a47a-ad8966d5087e","Type":"ContainerStarted","Data":"f87b12863adf8ed4999fe7348da7b1619198540ccdb079fedb1a3afe3af081b3"} Oct 08 14:19:37 crc kubenswrapper[4789]: I1008 14:19:37.904380 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d7c-account-create-cgqlh" event={"ID":"dbd72997-3d15-40ee-adfc-b1ffdf7667c9","Type":"ContainerStarted","Data":"efd39544356733c4b6909daa8d693bbe77fd4236b16e83b4cb22ea18e07587a0"} Oct 08 14:19:37 crc kubenswrapper[4789]: I1008 14:19:37.920650 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-7de8-account-create-4zc9n" podStartSLOduration=2.920630906 podStartE2EDuration="2.920630906s" podCreationTimestamp="2025-10-08 14:19:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:37.915164348 +0000 UTC m=+1117.821911840" watchObservedRunningTime="2025-10-08 14:19:37.920630906 +0000 UTC m=+1117.827378398" Oct 08 14:19:37 crc kubenswrapper[4789]: I1008 14:19:37.950115 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-2d7c-account-create-cgqlh" podStartSLOduration=5.950097085 podStartE2EDuration="5.950097085s" podCreationTimestamp="2025-10-08 14:19:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:37.935732375 +0000 UTC m=+1117.842479867" watchObservedRunningTime="2025-10-08 14:19:37.950097085 +0000 UTC m=+1117.856844577" Oct 08 14:19:37 crc kubenswrapper[4789]: I1008 14:19:37.953612 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-pd9bc" podStartSLOduration=5.95359878 podStartE2EDuration="5.95359878s" podCreationTimestamp="2025-10-08 14:19:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:37.951436891 +0000 UTC m=+1117.858184383" watchObservedRunningTime="2025-10-08 14:19:37.95359878 +0000 UTC m=+1117.860346272" Oct 08 14:19:38 crc kubenswrapper[4789]: I1008 14:19:38.235459 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:38 crc kubenswrapper[4789]: I1008 14:19:38.245196 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0462c12d-8df3-4734-9c2e-a925c9c06b9e-etc-swift\") pod \"swift-storage-0\" (UID: \"0462c12d-8df3-4734-9c2e-a925c9c06b9e\") " pod="openstack/swift-storage-0" Oct 08 14:19:38 crc kubenswrapper[4789]: I1008 14:19:38.400058 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 14:19:38 crc kubenswrapper[4789]: I1008 14:19:38.799444 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.110280 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.110532 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="prometheus" containerID="cri-o://bdd9e59a2c8f14973ba2ce56ccc6945b2ee66964a0e7da6de0f0e807a11f5d36" gracePeriod=600 Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.110611 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="thanos-sidecar" containerID="cri-o://0862d5c6b527f4d758ce54d7e6f4c60e4a85b8e64e1ccfdd9b63f944c7ef981c" gracePeriod=600 Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.110623 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="config-reloader" containerID="cri-o://134be9867b022fdfd415b2226c4968610d9f9cc7451cb4ad293311c57bc73b38" gracePeriod=600 Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.151677 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-notifications-server-0" podUID="55fc0a2c-48c9-4508-ae9d-5350f3298621" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.260015 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 14:19:39 crc kubenswrapper[4789]: W1008 14:19:39.261080 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0462c12d_8df3_4734_9c2e_a925c9c06b9e.slice/crio-5577b70053b4be13435aa93a6f320a19b5cc79b8662aef0a2ec70c1b0e9251ce WatchSource:0}: Error finding container 5577b70053b4be13435aa93a6f320a19b5cc79b8662aef0a2ec70c1b0e9251ce: Status 404 returned error can't find the container with id 5577b70053b4be13435aa93a6f320a19b5cc79b8662aef0a2ec70c1b0e9251ce Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.524976 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.110:5671: connect: connection refused" Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.921709 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"5577b70053b4be13435aa93a6f320a19b5cc79b8662aef0a2ec70c1b0e9251ce"} Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.925132 4789 generic.go:334] "Generic (PLEG): container finished" podID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerID="0862d5c6b527f4d758ce54d7e6f4c60e4a85b8e64e1ccfdd9b63f944c7ef981c" exitCode=0 Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.925189 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerDied","Data":"0862d5c6b527f4d758ce54d7e6f4c60e4a85b8e64e1ccfdd9b63f944c7ef981c"} Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.925227 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerDied","Data":"134be9867b022fdfd415b2226c4968610d9f9cc7451cb4ad293311c57bc73b38"} Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.925190 4789 generic.go:334] "Generic (PLEG): container finished" podID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerID="134be9867b022fdfd415b2226c4968610d9f9cc7451cb4ad293311c57bc73b38" exitCode=0 Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.925264 4789 generic.go:334] "Generic (PLEG): container finished" podID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerID="bdd9e59a2c8f14973ba2ce56ccc6945b2ee66964a0e7da6de0f0e807a11f5d36" exitCode=0 Oct 08 14:19:39 crc kubenswrapper[4789]: I1008 14:19:39.925284 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerDied","Data":"bdd9e59a2c8f14973ba2ce56ccc6945b2ee66964a0e7da6de0f0e807a11f5d36"} Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.938934 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.952752 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aaeb67a6-3e80-4d6e-9299-47634a566c72","Type":"ContainerDied","Data":"ffab64301e63b6cb5c0de92fcc0941f43ee102f740a4456bafd75b00b9a5b2a4"} Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.952801 4789 scope.go:117] "RemoveContainer" containerID="0862d5c6b527f4d758ce54d7e6f4c60e4a85b8e64e1ccfdd9b63f944c7ef981c" Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.952920 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.955454 4789 generic.go:334] "Generic (PLEG): container finished" podID="d3f0c4cf-3ebe-440d-a47a-ad8966d5087e" containerID="f87b12863adf8ed4999fe7348da7b1619198540ccdb079fedb1a3afe3af081b3" exitCode=0 Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.955525 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-7de8-account-create-4zc9n" event={"ID":"d3f0c4cf-3ebe-440d-a47a-ad8966d5087e","Type":"ContainerDied","Data":"f87b12863adf8ed4999fe7348da7b1619198540ccdb079fedb1a3afe3af081b3"} Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.965017 4789 generic.go:334] "Generic (PLEG): container finished" podID="dbd72997-3d15-40ee-adfc-b1ffdf7667c9" containerID="efd39544356733c4b6909daa8d693bbe77fd4236b16e83b4cb22ea18e07587a0" exitCode=0 Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.965167 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d7c-account-create-cgqlh" event={"ID":"dbd72997-3d15-40ee-adfc-b1ffdf7667c9","Type":"ContainerDied","Data":"efd39544356733c4b6909daa8d693bbe77fd4236b16e83b4cb22ea18e07587a0"} Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.971447 4789 generic.go:334] "Generic (PLEG): container finished" podID="bdab5d2d-97d5-4c68-ad3e-2589adfcadd4" containerID="e2a9d30875924c891b81d7e19dd1b2f459d6c0a4e8b702e049efaf70f4cc4f38" exitCode=0 Oct 08 14:19:40 crc kubenswrapper[4789]: I1008 14:19:40.971488 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pd9bc" event={"ID":"bdab5d2d-97d5-4c68-ad3e-2589adfcadd4","Type":"ContainerDied","Data":"e2a9d30875924c891b81d7e19dd1b2f459d6c0a4e8b702e049efaf70f4cc4f38"} Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.096039 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aaeb67a6-3e80-4d6e-9299-47634a566c72-prometheus-metric-storage-rulefiles-0\") pod \"aaeb67a6-3e80-4d6e-9299-47634a566c72\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.096099 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-tls-assets\") pod \"aaeb67a6-3e80-4d6e-9299-47634a566c72\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.096211 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-config\") pod \"aaeb67a6-3e80-4d6e-9299-47634a566c72\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.096258 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aaeb67a6-3e80-4d6e-9299-47634a566c72-config-out\") pod \"aaeb67a6-3e80-4d6e-9299-47634a566c72\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.096286 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-thanos-prometheus-http-client-file\") pod \"aaeb67a6-3e80-4d6e-9299-47634a566c72\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.096461 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"aaeb67a6-3e80-4d6e-9299-47634a566c72\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.096495 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-web-config\") pod \"aaeb67a6-3e80-4d6e-9299-47634a566c72\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.096563 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcm95\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-kube-api-access-rcm95\") pod \"aaeb67a6-3e80-4d6e-9299-47634a566c72\" (UID: \"aaeb67a6-3e80-4d6e-9299-47634a566c72\") " Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.098788 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaeb67a6-3e80-4d6e-9299-47634a566c72-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "aaeb67a6-3e80-4d6e-9299-47634a566c72" (UID: "aaeb67a6-3e80-4d6e-9299-47634a566c72"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.103290 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-kube-api-access-rcm95" (OuterVolumeSpecName: "kube-api-access-rcm95") pod "aaeb67a6-3e80-4d6e-9299-47634a566c72" (UID: "aaeb67a6-3e80-4d6e-9299-47634a566c72"). InnerVolumeSpecName "kube-api-access-rcm95". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.103405 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-config" (OuterVolumeSpecName: "config") pod "aaeb67a6-3e80-4d6e-9299-47634a566c72" (UID: "aaeb67a6-3e80-4d6e-9299-47634a566c72"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.108470 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "aaeb67a6-3e80-4d6e-9299-47634a566c72" (UID: "aaeb67a6-3e80-4d6e-9299-47634a566c72"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.110306 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaeb67a6-3e80-4d6e-9299-47634a566c72-config-out" (OuterVolumeSpecName: "config-out") pod "aaeb67a6-3e80-4d6e-9299-47634a566c72" (UID: "aaeb67a6-3e80-4d6e-9299-47634a566c72"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.126656 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "aaeb67a6-3e80-4d6e-9299-47634a566c72" (UID: "aaeb67a6-3e80-4d6e-9299-47634a566c72"). InnerVolumeSpecName "pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.127114 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "aaeb67a6-3e80-4d6e-9299-47634a566c72" (UID: "aaeb67a6-3e80-4d6e-9299-47634a566c72"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.167329 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-web-config" (OuterVolumeSpecName: "web-config") pod "aaeb67a6-3e80-4d6e-9299-47634a566c72" (UID: "aaeb67a6-3e80-4d6e-9299-47634a566c72"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.199770 4789 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.199858 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") on node \"crc\" " Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.199878 4789 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-web-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.199892 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcm95\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-kube-api-access-rcm95\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.199905 4789 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aaeb67a6-3e80-4d6e-9299-47634a566c72-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.199916 4789 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aaeb67a6-3e80-4d6e-9299-47634a566c72-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.199926 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/aaeb67a6-3e80-4d6e-9299-47634a566c72-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.199934 4789 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aaeb67a6-3e80-4d6e-9299-47634a566c72-config-out\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.233196 4789 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.233350 4789 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019") on node "crc" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.301288 4789 reconciler_common.go:293] "Volume detached for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.302522 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.313453 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.333324 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:19:41 crc kubenswrapper[4789]: E1008 14:19:41.333775 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="thanos-sidecar" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.333798 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="thanos-sidecar" Oct 08 14:19:41 crc kubenswrapper[4789]: E1008 14:19:41.333833 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="init-config-reloader" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.333843 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="init-config-reloader" Oct 08 14:19:41 crc kubenswrapper[4789]: E1008 14:19:41.333870 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="prometheus" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.333879 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="prometheus" Oct 08 14:19:41 crc kubenswrapper[4789]: E1008 14:19:41.333896 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="config-reloader" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.333905 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="config-reloader" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.334121 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="thanos-sidecar" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.337127 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="config-reloader" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.337179 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" containerName="prometheus" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.339378 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.349825 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.350033 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.350151 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.352367 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.353353 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-khr4j" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.353606 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.354047 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.360335 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.502222 4789 scope.go:117] "RemoveContainer" containerID="134be9867b022fdfd415b2226c4968610d9f9cc7451cb4ad293311c57bc73b38" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506009 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506056 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506097 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw6j4\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-kube-api-access-zw6j4\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506351 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506448 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0c9e486c-9ff9-45f3-860e-727852ae4461-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506531 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506563 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506590 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c9e486c-9ff9-45f3-860e-727852ae4461-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506616 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506653 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.506676 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.547840 4789 scope.go:117] "RemoveContainer" containerID="bdd9e59a2c8f14973ba2ce56ccc6945b2ee66964a0e7da6de0f0e807a11f5d36" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.568691 4789 scope.go:117] "RemoveContainer" containerID="468c5cd4b0553b344bb90ce3298ca311aa563ba304e0d98f2b97be854f4efbc1" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608086 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c9e486c-9ff9-45f3-860e-727852ae4461-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608151 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608190 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608212 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608267 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608309 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608436 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw6j4\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-kube-api-access-zw6j4\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608519 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608559 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0c9e486c-9ff9-45f3-860e-727852ae4461-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608603 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.608627 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.614548 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0c9e486c-9ff9-45f3-860e-727852ae4461-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.616583 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.616654 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.618282 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.618448 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.619691 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c9e486c-9ff9-45f3-860e-727852ae4461-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.619717 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.619920 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.621139 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.622581 4789 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.622616 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f75c8fbff2adebf98b3c9ad4427807236950b84a4f856fba0192ffb272b3d6c1/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.635677 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw6j4\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-kube-api-access-zw6j4\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.670256 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.964599 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.992197 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"28b15b9d74bfbaaaee37052db0c3fe61b9a66d8a8fe10726faf4cc6d22fe521c"} Oct 08 14:19:41 crc kubenswrapper[4789]: I1008 14:19:41.992250 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"6fef6c2cfe90553c46b576fdf7d807f69f7471225edfe9be3c98868af3896f0d"} Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.381597 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-7de8-account-create-4zc9n" Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.516385 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pd9bc" Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.529934 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcsvp\" (UniqueName: \"kubernetes.io/projected/d3f0c4cf-3ebe-440d-a47a-ad8966d5087e-kube-api-access-dcsvp\") pod \"d3f0c4cf-3ebe-440d-a47a-ad8966d5087e\" (UID: \"d3f0c4cf-3ebe-440d-a47a-ad8966d5087e\") " Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.540066 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f0c4cf-3ebe-440d-a47a-ad8966d5087e-kube-api-access-dcsvp" (OuterVolumeSpecName: "kube-api-access-dcsvp") pod "d3f0c4cf-3ebe-440d-a47a-ad8966d5087e" (UID: "d3f0c4cf-3ebe-440d-a47a-ad8966d5087e"). InnerVolumeSpecName "kube-api-access-dcsvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.585121 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d7c-account-create-cgqlh" Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.631128 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m79c\" (UniqueName: \"kubernetes.io/projected/bdab5d2d-97d5-4c68-ad3e-2589adfcadd4-kube-api-access-2m79c\") pod \"bdab5d2d-97d5-4c68-ad3e-2589adfcadd4\" (UID: \"bdab5d2d-97d5-4c68-ad3e-2589adfcadd4\") " Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.631467 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcsvp\" (UniqueName: \"kubernetes.io/projected/d3f0c4cf-3ebe-440d-a47a-ad8966d5087e-kube-api-access-dcsvp\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.637211 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdab5d2d-97d5-4c68-ad3e-2589adfcadd4-kube-api-access-2m79c" (OuterVolumeSpecName: "kube-api-access-2m79c") pod "bdab5d2d-97d5-4c68-ad3e-2589adfcadd4" (UID: "bdab5d2d-97d5-4c68-ad3e-2589adfcadd4"). InnerVolumeSpecName "kube-api-access-2m79c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.734556 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv6wr\" (UniqueName: \"kubernetes.io/projected/dbd72997-3d15-40ee-adfc-b1ffdf7667c9-kube-api-access-jv6wr\") pod \"dbd72997-3d15-40ee-adfc-b1ffdf7667c9\" (UID: \"dbd72997-3d15-40ee-adfc-b1ffdf7667c9\") " Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.734902 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m79c\" (UniqueName: \"kubernetes.io/projected/bdab5d2d-97d5-4c68-ad3e-2589adfcadd4-kube-api-access-2m79c\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.738031 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbd72997-3d15-40ee-adfc-b1ffdf7667c9-kube-api-access-jv6wr" (OuterVolumeSpecName: "kube-api-access-jv6wr") pod "dbd72997-3d15-40ee-adfc-b1ffdf7667c9" (UID: "dbd72997-3d15-40ee-adfc-b1ffdf7667c9"). InnerVolumeSpecName "kube-api-access-jv6wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.742356 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaeb67a6-3e80-4d6e-9299-47634a566c72" path="/var/lib/kubelet/pods/aaeb67a6-3e80-4d6e-9299-47634a566c72/volumes" Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.761758 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:19:42 crc kubenswrapper[4789]: I1008 14:19:42.837374 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv6wr\" (UniqueName: \"kubernetes.io/projected/dbd72997-3d15-40ee-adfc-b1ffdf7667c9-kube-api-access-jv6wr\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.016659 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-pd9bc" event={"ID":"bdab5d2d-97d5-4c68-ad3e-2589adfcadd4","Type":"ContainerDied","Data":"d0682ae32a29e35bafc73aa23f74ee29be604f5cd1f1807bd91a0c1752bb9503"} Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.016708 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0682ae32a29e35bafc73aa23f74ee29be604f5cd1f1807bd91a0c1752bb9503" Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.016814 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-pd9bc" Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.024579 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-7de8-account-create-4zc9n" event={"ID":"d3f0c4cf-3ebe-440d-a47a-ad8966d5087e","Type":"ContainerDied","Data":"f56ec899cacdd3a017d2dd4d6cee4ccfcf99f8e1fb40ab6d642f97ce307aa2b5"} Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.024624 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f56ec899cacdd3a017d2dd4d6cee4ccfcf99f8e1fb40ab6d642f97ce307aa2b5" Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.024602 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-7de8-account-create-4zc9n" Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.034051 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d7c-account-create-cgqlh" event={"ID":"dbd72997-3d15-40ee-adfc-b1ffdf7667c9","Type":"ContainerDied","Data":"91d94b4b082eab6d7ad0744082975e0ad80411542953186d636080a1b2e89707"} Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.034190 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91d94b4b082eab6d7ad0744082975e0ad80411542953186d636080a1b2e89707" Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.034753 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d7c-account-create-cgqlh" Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.037228 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerStarted","Data":"40eb210931b29358f97772f27f14fe98d53668bdeebb2e745fa80d2f4990d396"} Oct 08 14:19:43 crc kubenswrapper[4789]: I1008 14:19:43.049076 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"6bc608c6468d49438aa0c8e96ccd3fad804887c9c26a3c3fe3e519e342c81eaa"} Oct 08 14:19:44 crc kubenswrapper[4789]: I1008 14:19:44.060295 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"56682a28d9184d66d92873e17ae3faf37f0c499ab7c43ed1c99822a105fce731"} Oct 08 14:19:45 crc kubenswrapper[4789]: I1008 14:19:45.071188 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"cc11fb64c77ad435eef1c500371261eb41080baec2a2999f166f9b4d0374360f"} Oct 08 14:19:45 crc kubenswrapper[4789]: I1008 14:19:45.071483 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"52057e2e6015e71a7094ef9ccab666dc313c7f3753ab037a5bc5617e6386ecb7"} Oct 08 14:19:45 crc kubenswrapper[4789]: I1008 14:19:45.071496 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"20dc17fc5e5ef80998051652577098552da8e7138ae219a9514561225fb3b6b8"} Oct 08 14:19:46 crc kubenswrapper[4789]: I1008 14:19:46.106480 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerStarted","Data":"f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098"} Oct 08 14:19:46 crc kubenswrapper[4789]: I1008 14:19:46.123680 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"73024d3323ac963368d0179f35c8d249d49bbbbb02c6a57235144223b9eb5b3a"} Oct 08 14:19:46 crc kubenswrapper[4789]: I1008 14:19:46.123730 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"135367f13421baad7816db1b9975deac76e3694664d9dc8786c0563ef9c93d35"} Oct 08 14:19:47 crc kubenswrapper[4789]: I1008 14:19:47.161166 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"fe76da9254d34d0b300490d62646fdad945791aa985d711606865c4a184bbc69"} Oct 08 14:19:47 crc kubenswrapper[4789]: I1008 14:19:47.162132 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"e2e51e00d1b80879ecff1d4a320fcb0be49f1d6027316de40a51a6252610a46d"} Oct 08 14:19:47 crc kubenswrapper[4789]: I1008 14:19:47.162231 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"2f066c422470c4ca0ac957186cf9a2ca22f89160b7c0cc8799da70a624b3aaa6"} Oct 08 14:19:47 crc kubenswrapper[4789]: I1008 14:19:47.162245 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"934e99c7408cbd0fb016c4c2f601a9f4449fac901dd3ef18b9c3a271c4d20c50"} Oct 08 14:19:47 crc kubenswrapper[4789]: I1008 14:19:47.162253 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"0fbddbfdacbc368d2ff163894be5e0c1d876181f1bcdf79b92ae3d99bba75216"} Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.188167 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0462c12d-8df3-4734-9c2e-a925c9c06b9e","Type":"ContainerStarted","Data":"7d363c70b59233aea99341f55776273f52c08e9cca9dad634c73d38a1ab9e72f"} Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.219699 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.619122692 podStartE2EDuration="43.219682481s" podCreationTimestamp="2025-10-08 14:19:05 +0000 UTC" firstStartedPulling="2025-10-08 14:19:39.264148269 +0000 UTC m=+1119.170895751" lastFinishedPulling="2025-10-08 14:19:45.864708048 +0000 UTC m=+1125.771455540" observedRunningTime="2025-10-08 14:19:48.215549769 +0000 UTC m=+1128.122297261" watchObservedRunningTime="2025-10-08 14:19:48.219682481 +0000 UTC m=+1128.126429973" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.479842 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fcb49bfb9-8ch52"] Oct 08 14:19:48 crc kubenswrapper[4789]: E1008 14:19:48.480246 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f0c4cf-3ebe-440d-a47a-ad8966d5087e" containerName="mariadb-account-create" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.480263 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f0c4cf-3ebe-440d-a47a-ad8966d5087e" containerName="mariadb-account-create" Oct 08 14:19:48 crc kubenswrapper[4789]: E1008 14:19:48.480284 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdab5d2d-97d5-4c68-ad3e-2589adfcadd4" containerName="mariadb-database-create" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.480289 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdab5d2d-97d5-4c68-ad3e-2589adfcadd4" containerName="mariadb-database-create" Oct 08 14:19:48 crc kubenswrapper[4789]: E1008 14:19:48.480298 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbd72997-3d15-40ee-adfc-b1ffdf7667c9" containerName="mariadb-account-create" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.480304 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbd72997-3d15-40ee-adfc-b1ffdf7667c9" containerName="mariadb-account-create" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.480464 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f0c4cf-3ebe-440d-a47a-ad8966d5087e" containerName="mariadb-account-create" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.480476 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbd72997-3d15-40ee-adfc-b1ffdf7667c9" containerName="mariadb-account-create" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.480497 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdab5d2d-97d5-4c68-ad3e-2589adfcadd4" containerName="mariadb-database-create" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.481434 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.484238 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.497156 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fcb49bfb9-8ch52"] Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.634546 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nd95\" (UniqueName: \"kubernetes.io/projected/bc8dc138-180a-44c7-817e-d3baf5231352-kube-api-access-9nd95\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.634637 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-nb\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.634839 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-swift-storage-0\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.634898 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-sb\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.634923 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-svc\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.635035 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-config\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.736606 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-sb\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.736650 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-svc\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.736697 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-config\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.736744 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nd95\" (UniqueName: \"kubernetes.io/projected/bc8dc138-180a-44c7-817e-d3baf5231352-kube-api-access-9nd95\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.736811 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-nb\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.736880 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-swift-storage-0\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.737674 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-config\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.737697 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-sb\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.737849 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-swift-storage-0\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.738069 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-svc\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.738395 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-nb\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.759953 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nd95\" (UniqueName: \"kubernetes.io/projected/bc8dc138-180a-44c7-817e-d3baf5231352-kube-api-access-9nd95\") pod \"dnsmasq-dns-5fcb49bfb9-8ch52\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.798469 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:48 crc kubenswrapper[4789]: I1008 14:19:48.798591 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 08 14:19:49 crc kubenswrapper[4789]: I1008 14:19:49.151224 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-notifications-server-0" podUID="55fc0a2c-48c9-4508-ae9d-5350f3298621" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.109:5671: connect: connection refused" Oct 08 14:19:49 crc kubenswrapper[4789]: I1008 14:19:49.262732 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fcb49bfb9-8ch52"] Oct 08 14:19:49 crc kubenswrapper[4789]: I1008 14:19:49.524318 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.110:5671: connect: connection refused" Oct 08 14:19:50 crc kubenswrapper[4789]: I1008 14:19:50.203215 4789 generic.go:334] "Generic (PLEG): container finished" podID="bc8dc138-180a-44c7-817e-d3baf5231352" containerID="0a1993ad5ef13630ccb232a1839ea051ad883fce38f6cfce771177079935277d" exitCode=0 Oct 08 14:19:50 crc kubenswrapper[4789]: I1008 14:19:50.203254 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" event={"ID":"bc8dc138-180a-44c7-817e-d3baf5231352","Type":"ContainerDied","Data":"0a1993ad5ef13630ccb232a1839ea051ad883fce38f6cfce771177079935277d"} Oct 08 14:19:50 crc kubenswrapper[4789]: I1008 14:19:50.203316 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" event={"ID":"bc8dc138-180a-44c7-817e-d3baf5231352","Type":"ContainerStarted","Data":"3664e4f95790cb3919859131f08344a1c933ca1d668c516796f7d01c71228077"} Oct 08 14:19:51 crc kubenswrapper[4789]: I1008 14:19:51.213635 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" event={"ID":"bc8dc138-180a-44c7-817e-d3baf5231352","Type":"ContainerStarted","Data":"4583edbf8887854b7713f005874fe01ccf08147dacbba0f7e06e974f3afae0b2"} Oct 08 14:19:51 crc kubenswrapper[4789]: I1008 14:19:51.214734 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:51 crc kubenswrapper[4789]: I1008 14:19:51.247323 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" podStartSLOduration=3.247306976 podStartE2EDuration="3.247306976s" podCreationTimestamp="2025-10-08 14:19:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:51.243774921 +0000 UTC m=+1131.150522513" watchObservedRunningTime="2025-10-08 14:19:51.247306976 +0000 UTC m=+1131.154054458" Oct 08 14:19:52 crc kubenswrapper[4789]: I1008 14:19:52.286348 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-26f3-account-create-4qzvp"] Oct 08 14:19:52 crc kubenswrapper[4789]: I1008 14:19:52.287362 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-26f3-account-create-4qzvp" Oct 08 14:19:52 crc kubenswrapper[4789]: I1008 14:19:52.289930 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 08 14:19:52 crc kubenswrapper[4789]: I1008 14:19:52.301484 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-26f3-account-create-4qzvp"] Oct 08 14:19:52 crc kubenswrapper[4789]: I1008 14:19:52.405318 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqbhw\" (UniqueName: \"kubernetes.io/projected/d43aacf0-f8aa-483b-a0dc-a5b69f9c2984-kube-api-access-tqbhw\") pod \"keystone-26f3-account-create-4qzvp\" (UID: \"d43aacf0-f8aa-483b-a0dc-a5b69f9c2984\") " pod="openstack/keystone-26f3-account-create-4qzvp" Oct 08 14:19:52 crc kubenswrapper[4789]: I1008 14:19:52.507138 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqbhw\" (UniqueName: \"kubernetes.io/projected/d43aacf0-f8aa-483b-a0dc-a5b69f9c2984-kube-api-access-tqbhw\") pod \"keystone-26f3-account-create-4qzvp\" (UID: \"d43aacf0-f8aa-483b-a0dc-a5b69f9c2984\") " pod="openstack/keystone-26f3-account-create-4qzvp" Oct 08 14:19:52 crc kubenswrapper[4789]: I1008 14:19:52.535178 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqbhw\" (UniqueName: \"kubernetes.io/projected/d43aacf0-f8aa-483b-a0dc-a5b69f9c2984-kube-api-access-tqbhw\") pod \"keystone-26f3-account-create-4qzvp\" (UID: \"d43aacf0-f8aa-483b-a0dc-a5b69f9c2984\") " pod="openstack/keystone-26f3-account-create-4qzvp" Oct 08 14:19:52 crc kubenswrapper[4789]: I1008 14:19:52.603619 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-26f3-account-create-4qzvp" Oct 08 14:19:53 crc kubenswrapper[4789]: I1008 14:19:53.000910 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-26f3-account-create-4qzvp"] Oct 08 14:19:53 crc kubenswrapper[4789]: I1008 14:19:53.231597 4789 generic.go:334] "Generic (PLEG): container finished" podID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerID="f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098" exitCode=0 Oct 08 14:19:53 crc kubenswrapper[4789]: I1008 14:19:53.231660 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerDied","Data":"f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098"} Oct 08 14:19:53 crc kubenswrapper[4789]: I1008 14:19:53.235033 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-26f3-account-create-4qzvp" event={"ID":"d43aacf0-f8aa-483b-a0dc-a5b69f9c2984","Type":"ContainerStarted","Data":"f82778926ec6c715ebb40417240a144d7ec60fd891af6bd8279674431ed4e457"} Oct 08 14:19:54 crc kubenswrapper[4789]: I1008 14:19:54.247954 4789 generic.go:334] "Generic (PLEG): container finished" podID="d43aacf0-f8aa-483b-a0dc-a5b69f9c2984" containerID="905c5315d8f00d0d37d0da6d4ca72b638315f50d6e6f45fa0a0b31a0f3c00ed4" exitCode=0 Oct 08 14:19:54 crc kubenswrapper[4789]: I1008 14:19:54.248032 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-26f3-account-create-4qzvp" event={"ID":"d43aacf0-f8aa-483b-a0dc-a5b69f9c2984","Type":"ContainerDied","Data":"905c5315d8f00d0d37d0da6d4ca72b638315f50d6e6f45fa0a0b31a0f3c00ed4"} Oct 08 14:19:54 crc kubenswrapper[4789]: I1008 14:19:54.251233 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerStarted","Data":"4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705"} Oct 08 14:19:55 crc kubenswrapper[4789]: I1008 14:19:55.591128 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-26f3-account-create-4qzvp" Oct 08 14:19:55 crc kubenswrapper[4789]: I1008 14:19:55.656269 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqbhw\" (UniqueName: \"kubernetes.io/projected/d43aacf0-f8aa-483b-a0dc-a5b69f9c2984-kube-api-access-tqbhw\") pod \"d43aacf0-f8aa-483b-a0dc-a5b69f9c2984\" (UID: \"d43aacf0-f8aa-483b-a0dc-a5b69f9c2984\") " Oct 08 14:19:55 crc kubenswrapper[4789]: I1008 14:19:55.664852 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d43aacf0-f8aa-483b-a0dc-a5b69f9c2984-kube-api-access-tqbhw" (OuterVolumeSpecName: "kube-api-access-tqbhw") pod "d43aacf0-f8aa-483b-a0dc-a5b69f9c2984" (UID: "d43aacf0-f8aa-483b-a0dc-a5b69f9c2984"). InnerVolumeSpecName "kube-api-access-tqbhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:55 crc kubenswrapper[4789]: I1008 14:19:55.758022 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqbhw\" (UniqueName: \"kubernetes.io/projected/d43aacf0-f8aa-483b-a0dc-a5b69f9c2984-kube-api-access-tqbhw\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:56 crc kubenswrapper[4789]: I1008 14:19:56.266118 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-26f3-account-create-4qzvp" event={"ID":"d43aacf0-f8aa-483b-a0dc-a5b69f9c2984","Type":"ContainerDied","Data":"f82778926ec6c715ebb40417240a144d7ec60fd891af6bd8279674431ed4e457"} Oct 08 14:19:56 crc kubenswrapper[4789]: I1008 14:19:56.266155 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-26f3-account-create-4qzvp" Oct 08 14:19:56 crc kubenswrapper[4789]: I1008 14:19:56.266169 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f82778926ec6c715ebb40417240a144d7ec60fd891af6bd8279674431ed4e457" Oct 08 14:19:56 crc kubenswrapper[4789]: I1008 14:19:56.433348 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:19:56 crc kubenswrapper[4789]: I1008 14:19:56.433418 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:19:57 crc kubenswrapper[4789]: I1008 14:19:57.276782 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerStarted","Data":"0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465"} Oct 08 14:19:57 crc kubenswrapper[4789]: I1008 14:19:57.276830 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerStarted","Data":"c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c"} Oct 08 14:19:57 crc kubenswrapper[4789]: I1008 14:19:57.302937 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.302917824 podStartE2EDuration="16.302917824s" podCreationTimestamp="2025-10-08 14:19:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:19:57.30167457 +0000 UTC m=+1137.208422082" watchObservedRunningTime="2025-10-08 14:19:57.302917824 +0000 UTC m=+1137.209665316" Oct 08 14:19:58 crc kubenswrapper[4789]: I1008 14:19:58.799179 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:19:58 crc kubenswrapper[4789]: I1008 14:19:58.800153 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:19:58 crc kubenswrapper[4789]: I1008 14:19:58.893286 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58f55f8967-2cng7"] Oct 08 14:19:58 crc kubenswrapper[4789]: I1008 14:19:58.893638 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" podUID="1b0c0d62-4dc5-49af-bd91-746237228359" containerName="dnsmasq-dns" containerID="cri-o://e187ba8bfc6f6962c902db0c44f12c868ae0283bb53c8d1f44a15ca43db752c3" gracePeriod=10 Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.155726 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-notifications-server-0" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.312378 4789 generic.go:334] "Generic (PLEG): container finished" podID="1b0c0d62-4dc5-49af-bd91-746237228359" containerID="e187ba8bfc6f6962c902db0c44f12c868ae0283bb53c8d1f44a15ca43db752c3" exitCode=0 Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.312421 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" event={"ID":"1b0c0d62-4dc5-49af-bd91-746237228359","Type":"ContainerDied","Data":"e187ba8bfc6f6962c902db0c44f12c868ae0283bb53c8d1f44a15ca43db752c3"} Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.384569 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.424321 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-dns-svc\") pod \"1b0c0d62-4dc5-49af-bd91-746237228359\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.424413 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-sb\") pod \"1b0c0d62-4dc5-49af-bd91-746237228359\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.424494 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbvrg\" (UniqueName: \"kubernetes.io/projected/1b0c0d62-4dc5-49af-bd91-746237228359-kube-api-access-jbvrg\") pod \"1b0c0d62-4dc5-49af-bd91-746237228359\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.424641 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-nb\") pod \"1b0c0d62-4dc5-49af-bd91-746237228359\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.424671 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-config\") pod \"1b0c0d62-4dc5-49af-bd91-746237228359\" (UID: \"1b0c0d62-4dc5-49af-bd91-746237228359\") " Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.432044 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b0c0d62-4dc5-49af-bd91-746237228359-kube-api-access-jbvrg" (OuterVolumeSpecName: "kube-api-access-jbvrg") pod "1b0c0d62-4dc5-49af-bd91-746237228359" (UID: "1b0c0d62-4dc5-49af-bd91-746237228359"). InnerVolumeSpecName "kube-api-access-jbvrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.477811 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-config" (OuterVolumeSpecName: "config") pod "1b0c0d62-4dc5-49af-bd91-746237228359" (UID: "1b0c0d62-4dc5-49af-bd91-746237228359"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.480575 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b0c0d62-4dc5-49af-bd91-746237228359" (UID: "1b0c0d62-4dc5-49af-bd91-746237228359"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.482666 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b0c0d62-4dc5-49af-bd91-746237228359" (UID: "1b0c0d62-4dc5-49af-bd91-746237228359"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.487792 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b0c0d62-4dc5-49af-bd91-746237228359" (UID: "1b0c0d62-4dc5-49af-bd91-746237228359"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.526576 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.527054 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.527094 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.527110 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.527125 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b0c0d62-4dc5-49af-bd91-746237228359-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:19:59 crc kubenswrapper[4789]: I1008 14:19:59.527137 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbvrg\" (UniqueName: \"kubernetes.io/projected/1b0c0d62-4dc5-49af-bd91-746237228359-kube-api-access-jbvrg\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.256397 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-xjfkl"] Oct 08 14:20:00 crc kubenswrapper[4789]: E1008 14:20:00.256997 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b0c0d62-4dc5-49af-bd91-746237228359" containerName="init" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.257008 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b0c0d62-4dc5-49af-bd91-746237228359" containerName="init" Oct 08 14:20:00 crc kubenswrapper[4789]: E1008 14:20:00.257019 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b0c0d62-4dc5-49af-bd91-746237228359" containerName="dnsmasq-dns" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.257025 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b0c0d62-4dc5-49af-bd91-746237228359" containerName="dnsmasq-dns" Oct 08 14:20:00 crc kubenswrapper[4789]: E1008 14:20:00.257042 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d43aacf0-f8aa-483b-a0dc-a5b69f9c2984" containerName="mariadb-account-create" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.257049 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d43aacf0-f8aa-483b-a0dc-a5b69f9c2984" containerName="mariadb-account-create" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.257264 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b0c0d62-4dc5-49af-bd91-746237228359" containerName="dnsmasq-dns" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.257282 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d43aacf0-f8aa-483b-a0dc-a5b69f9c2984" containerName="mariadb-account-create" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.257953 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xjfkl" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.265235 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xjfkl"] Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.322096 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" event={"ID":"1b0c0d62-4dc5-49af-bd91-746237228359","Type":"ContainerDied","Data":"d3fa09c60e4b27d7623f47097e723bae79dc17330c28ccfbd26bf515d0d61548"} Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.322146 4789 scope.go:117] "RemoveContainer" containerID="e187ba8bfc6f6962c902db0c44f12c868ae0283bb53c8d1f44a15ca43db752c3" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.322175 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58f55f8967-2cng7" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.339360 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9phkf\" (UniqueName: \"kubernetes.io/projected/48f2731e-7573-4376-96bb-e19ac616a706-kube-api-access-9phkf\") pod \"glance-db-create-xjfkl\" (UID: \"48f2731e-7573-4376-96bb-e19ac616a706\") " pod="openstack/glance-db-create-xjfkl" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.342186 4789 scope.go:117] "RemoveContainer" containerID="43ca0225ac870080925935c167059e44f176fab94ada40fbe3c6a60fb950fa6a" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.352125 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58f55f8967-2cng7"] Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.359959 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58f55f8967-2cng7"] Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.441283 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9phkf\" (UniqueName: \"kubernetes.io/projected/48f2731e-7573-4376-96bb-e19ac616a706-kube-api-access-9phkf\") pod \"glance-db-create-xjfkl\" (UID: \"48f2731e-7573-4376-96bb-e19ac616a706\") " pod="openstack/glance-db-create-xjfkl" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.457858 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9phkf\" (UniqueName: \"kubernetes.io/projected/48f2731e-7573-4376-96bb-e19ac616a706-kube-api-access-9phkf\") pod \"glance-db-create-xjfkl\" (UID: \"48f2731e-7573-4376-96bb-e19ac616a706\") " pod="openstack/glance-db-create-xjfkl" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.572291 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xjfkl" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.765967 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b0c0d62-4dc5-49af-bd91-746237228359" path="/var/lib/kubelet/pods/1b0c0d62-4dc5-49af-bd91-746237228359/volumes" Oct 08 14:20:00 crc kubenswrapper[4789]: I1008 14:20:00.952641 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xjfkl"] Oct 08 14:20:00 crc kubenswrapper[4789]: W1008 14:20:00.972308 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48f2731e_7573_4376_96bb_e19ac616a706.slice/crio-0964800637d49c3c79dd37de820284c053c94912051a3d971ce16282b9d66bcd WatchSource:0}: Error finding container 0964800637d49c3c79dd37de820284c053c94912051a3d971ce16282b9d66bcd: Status 404 returned error can't find the container with id 0964800637d49c3c79dd37de820284c053c94912051a3d971ce16282b9d66bcd Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.333762 4789 generic.go:334] "Generic (PLEG): container finished" podID="48f2731e-7573-4376-96bb-e19ac616a706" containerID="7c5cab1656774270d9ae7b1c3f873d33d7a50c423794b564e1e73089bf5a9731" exitCode=0 Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.334125 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xjfkl" event={"ID":"48f2731e-7573-4376-96bb-e19ac616a706","Type":"ContainerDied","Data":"7c5cab1656774270d9ae7b1c3f873d33d7a50c423794b564e1e73089bf5a9731"} Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.334183 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xjfkl" event={"ID":"48f2731e-7573-4376-96bb-e19ac616a706","Type":"ContainerStarted","Data":"0964800637d49c3c79dd37de820284c053c94912051a3d971ce16282b9d66bcd"} Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.353736 4789 scope.go:117] "RemoveContainer" containerID="f3e020da59405436a0f3d63d89a3ea41e5db79a365c74a7154d9d7ca41442840" Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.718187 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4fhfk"] Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.719611 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4fhfk" Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.724349 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4fhfk"] Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.770147 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g94mk\" (UniqueName: \"kubernetes.io/projected/8e2b82e2-a807-49f8-9c77-f12a8b8bde3c-kube-api-access-g94mk\") pod \"barbican-db-create-4fhfk\" (UID: \"8e2b82e2-a807-49f8-9c77-f12a8b8bde3c\") " pod="openstack/barbican-db-create-4fhfk" Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.872289 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g94mk\" (UniqueName: \"kubernetes.io/projected/8e2b82e2-a807-49f8-9c77-f12a8b8bde3c-kube-api-access-g94mk\") pod \"barbican-db-create-4fhfk\" (UID: \"8e2b82e2-a807-49f8-9c77-f12a8b8bde3c\") " pod="openstack/barbican-db-create-4fhfk" Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.895825 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g94mk\" (UniqueName: \"kubernetes.io/projected/8e2b82e2-a807-49f8-9c77-f12a8b8bde3c-kube-api-access-g94mk\") pod \"barbican-db-create-4fhfk\" (UID: \"8e2b82e2-a807-49f8-9c77-f12a8b8bde3c\") " pod="openstack/barbican-db-create-4fhfk" Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.919062 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-9wtgk"] Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.920149 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9wtgk" Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.929675 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9wtgk"] Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.965002 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 08 14:20:01 crc kubenswrapper[4789]: I1008 14:20:01.973865 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt28n\" (UniqueName: \"kubernetes.io/projected/19ded6af-86ab-4942-b141-14db479a5986-kube-api-access-jt28n\") pod \"cinder-db-create-9wtgk\" (UID: \"19ded6af-86ab-4942-b141-14db479a5986\") " pod="openstack/cinder-db-create-9wtgk" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.036376 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4fhfk" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.059219 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-6w7kw"] Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.060642 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.063464 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-j7bbz" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.063504 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.074440 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-6w7kw"] Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.084787 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-db-sync-config-data\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.084930 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt28n\" (UniqueName: \"kubernetes.io/projected/19ded6af-86ab-4942-b141-14db479a5986-kube-api-access-jt28n\") pod \"cinder-db-create-9wtgk\" (UID: \"19ded6af-86ab-4942-b141-14db479a5986\") " pod="openstack/cinder-db-create-9wtgk" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.084963 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-config-data\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.085026 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-combined-ca-bundle\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.085194 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vb5x\" (UniqueName: \"kubernetes.io/projected/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-kube-api-access-5vb5x\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.119316 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt28n\" (UniqueName: \"kubernetes.io/projected/19ded6af-86ab-4942-b141-14db479a5986-kube-api-access-jt28n\") pod \"cinder-db-create-9wtgk\" (UID: \"19ded6af-86ab-4942-b141-14db479a5986\") " pod="openstack/cinder-db-create-9wtgk" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.128179 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-7nqh5"] Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.129241 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7nqh5" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.143227 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7nqh5"] Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.188541 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-db-sync-config-data\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.188602 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-config-data\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.188627 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nt7s\" (UniqueName: \"kubernetes.io/projected/a038fc03-8ce7-4641-a8f1-1e2a748847a9-kube-api-access-4nt7s\") pod \"neutron-db-create-7nqh5\" (UID: \"a038fc03-8ce7-4641-a8f1-1e2a748847a9\") " pod="openstack/neutron-db-create-7nqh5" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.188648 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-combined-ca-bundle\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.188711 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vb5x\" (UniqueName: \"kubernetes.io/projected/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-kube-api-access-5vb5x\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.197837 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-config-data\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.205236 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-combined-ca-bundle\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.208681 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-db-sync-config-data\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.213193 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vb5x\" (UniqueName: \"kubernetes.io/projected/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-kube-api-access-5vb5x\") pod \"watcher-db-sync-6w7kw\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.214531 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-srh88"] Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.215518 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.220806 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.220979 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.220806 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.223255 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8h2vv" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.254881 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9wtgk" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.256056 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-srh88"] Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.290381 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-486gf\" (UniqueName: \"kubernetes.io/projected/187c395c-7b8d-45e7-b96c-a93cea38b025-kube-api-access-486gf\") pod \"keystone-db-sync-srh88\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.290431 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-combined-ca-bundle\") pod \"keystone-db-sync-srh88\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.290466 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nt7s\" (UniqueName: \"kubernetes.io/projected/a038fc03-8ce7-4641-a8f1-1e2a748847a9-kube-api-access-4nt7s\") pod \"neutron-db-create-7nqh5\" (UID: \"a038fc03-8ce7-4641-a8f1-1e2a748847a9\") " pod="openstack/neutron-db-create-7nqh5" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.290621 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-config-data\") pod \"keystone-db-sync-srh88\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.365921 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nt7s\" (UniqueName: \"kubernetes.io/projected/a038fc03-8ce7-4641-a8f1-1e2a748847a9-kube-api-access-4nt7s\") pod \"neutron-db-create-7nqh5\" (UID: \"a038fc03-8ce7-4641-a8f1-1e2a748847a9\") " pod="openstack/neutron-db-create-7nqh5" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.388539 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.392491 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-config-data\") pod \"keystone-db-sync-srh88\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.392615 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-486gf\" (UniqueName: \"kubernetes.io/projected/187c395c-7b8d-45e7-b96c-a93cea38b025-kube-api-access-486gf\") pod \"keystone-db-sync-srh88\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.392643 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-combined-ca-bundle\") pod \"keystone-db-sync-srh88\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.396929 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-combined-ca-bundle\") pod \"keystone-db-sync-srh88\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.397768 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-config-data\") pod \"keystone-db-sync-srh88\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.422404 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-486gf\" (UniqueName: \"kubernetes.io/projected/187c395c-7b8d-45e7-b96c-a93cea38b025-kube-api-access-486gf\") pod \"keystone-db-sync-srh88\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.490638 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7nqh5" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.646414 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.704371 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4fhfk"] Oct 08 14:20:02 crc kubenswrapper[4789]: W1008 14:20:02.705012 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e2b82e2_a807_49f8_9c77_f12a8b8bde3c.slice/crio-5ba2d8c54d2c30ff5993f40827928ce743434b67c024cc55650a86a6d2005d00 WatchSource:0}: Error finding container 5ba2d8c54d2c30ff5993f40827928ce743434b67c024cc55650a86a6d2005d00: Status 404 returned error can't find the container with id 5ba2d8c54d2c30ff5993f40827928ce743434b67c024cc55650a86a6d2005d00 Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.825004 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xjfkl" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.901978 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9phkf\" (UniqueName: \"kubernetes.io/projected/48f2731e-7573-4376-96bb-e19ac616a706-kube-api-access-9phkf\") pod \"48f2731e-7573-4376-96bb-e19ac616a706\" (UID: \"48f2731e-7573-4376-96bb-e19ac616a706\") " Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.907183 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48f2731e-7573-4376-96bb-e19ac616a706-kube-api-access-9phkf" (OuterVolumeSpecName: "kube-api-access-9phkf") pod "48f2731e-7573-4376-96bb-e19ac616a706" (UID: "48f2731e-7573-4376-96bb-e19ac616a706"). InnerVolumeSpecName "kube-api-access-9phkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:02 crc kubenswrapper[4789]: I1008 14:20:02.916786 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-9wtgk"] Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.010117 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9phkf\" (UniqueName: \"kubernetes.io/projected/48f2731e-7573-4376-96bb-e19ac616a706-kube-api-access-9phkf\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.066123 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-6w7kw"] Oct 08 14:20:03 crc kubenswrapper[4789]: W1008 14:20:03.129403 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc629d8f6_586d_4f5c_a7ea_2e6912a02b03.slice/crio-841174b907edcbd61238e44f43e55b9fe6772f86e66ed9572539dde5b384f318 WatchSource:0}: Error finding container 841174b907edcbd61238e44f43e55b9fe6772f86e66ed9572539dde5b384f318: Status 404 returned error can't find the container with id 841174b907edcbd61238e44f43e55b9fe6772f86e66ed9572539dde5b384f318 Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.213862 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7nqh5"] Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.319314 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-srh88"] Oct 08 14:20:03 crc kubenswrapper[4789]: W1008 14:20:03.323654 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod187c395c_7b8d_45e7_b96c_a93cea38b025.slice/crio-b7d97602be651ccee34c51aeb9f5a8e445064a73c4596a2ce4ec5718f2174b47 WatchSource:0}: Error finding container b7d97602be651ccee34c51aeb9f5a8e445064a73c4596a2ce4ec5718f2174b47: Status 404 returned error can't find the container with id b7d97602be651ccee34c51aeb9f5a8e445064a73c4596a2ce4ec5718f2174b47 Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.371927 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9wtgk" event={"ID":"19ded6af-86ab-4942-b141-14db479a5986","Type":"ContainerStarted","Data":"2362393d2ec81bda1d6fb6b193b12618151d9c80f38bc6f937db8e07b6acdd1b"} Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.371973 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9wtgk" event={"ID":"19ded6af-86ab-4942-b141-14db479a5986","Type":"ContainerStarted","Data":"dc841c4113cdee897526ec57382e95ee7bcdb6e2f9e411201c6c3b39c2358bfb"} Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.376091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-6w7kw" event={"ID":"c629d8f6-586d-4f5c-a7ea-2e6912a02b03","Type":"ContainerStarted","Data":"841174b907edcbd61238e44f43e55b9fe6772f86e66ed9572539dde5b384f318"} Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.377157 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-srh88" event={"ID":"187c395c-7b8d-45e7-b96c-a93cea38b025","Type":"ContainerStarted","Data":"b7d97602be651ccee34c51aeb9f5a8e445064a73c4596a2ce4ec5718f2174b47"} Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.378307 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xjfkl" event={"ID":"48f2731e-7573-4376-96bb-e19ac616a706","Type":"ContainerDied","Data":"0964800637d49c3c79dd37de820284c053c94912051a3d971ce16282b9d66bcd"} Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.378332 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0964800637d49c3c79dd37de820284c053c94912051a3d971ce16282b9d66bcd" Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.378374 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xjfkl" Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.381567 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7nqh5" event={"ID":"a038fc03-8ce7-4641-a8f1-1e2a748847a9","Type":"ContainerStarted","Data":"32ce0c0e50306a38aa8236b25984940d34f526946de41510443f36af82896825"} Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.382850 4789 generic.go:334] "Generic (PLEG): container finished" podID="8e2b82e2-a807-49f8-9c77-f12a8b8bde3c" containerID="2e3aa0666eef15cf20b39327a33e2f1c526f956d6c495f0f19590c680f861dc4" exitCode=0 Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.382876 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4fhfk" event={"ID":"8e2b82e2-a807-49f8-9c77-f12a8b8bde3c","Type":"ContainerDied","Data":"2e3aa0666eef15cf20b39327a33e2f1c526f956d6c495f0f19590c680f861dc4"} Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.382891 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4fhfk" event={"ID":"8e2b82e2-a807-49f8-9c77-f12a8b8bde3c","Type":"ContainerStarted","Data":"5ba2d8c54d2c30ff5993f40827928ce743434b67c024cc55650a86a6d2005d00"} Oct 08 14:20:03 crc kubenswrapper[4789]: I1008 14:20:03.395457 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-9wtgk" podStartSLOduration=2.395437782 podStartE2EDuration="2.395437782s" podCreationTimestamp="2025-10-08 14:20:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:20:03.39058347 +0000 UTC m=+1143.297330962" watchObservedRunningTime="2025-10-08 14:20:03.395437782 +0000 UTC m=+1143.302185274" Oct 08 14:20:04 crc kubenswrapper[4789]: I1008 14:20:04.395317 4789 generic.go:334] "Generic (PLEG): container finished" podID="a038fc03-8ce7-4641-a8f1-1e2a748847a9" containerID="8ad7e6f23e8f8181d1905ac0b159f23d29a075cb04967e012a3e50d3819aba18" exitCode=0 Oct 08 14:20:04 crc kubenswrapper[4789]: I1008 14:20:04.396056 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7nqh5" event={"ID":"a038fc03-8ce7-4641-a8f1-1e2a748847a9","Type":"ContainerDied","Data":"8ad7e6f23e8f8181d1905ac0b159f23d29a075cb04967e012a3e50d3819aba18"} Oct 08 14:20:04 crc kubenswrapper[4789]: I1008 14:20:04.398610 4789 generic.go:334] "Generic (PLEG): container finished" podID="19ded6af-86ab-4942-b141-14db479a5986" containerID="2362393d2ec81bda1d6fb6b193b12618151d9c80f38bc6f937db8e07b6acdd1b" exitCode=0 Oct 08 14:20:04 crc kubenswrapper[4789]: I1008 14:20:04.398793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9wtgk" event={"ID":"19ded6af-86ab-4942-b141-14db479a5986","Type":"ContainerDied","Data":"2362393d2ec81bda1d6fb6b193b12618151d9c80f38bc6f937db8e07b6acdd1b"} Oct 08 14:20:04 crc kubenswrapper[4789]: I1008 14:20:04.873072 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4fhfk" Oct 08 14:20:04 crc kubenswrapper[4789]: I1008 14:20:04.952633 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g94mk\" (UniqueName: \"kubernetes.io/projected/8e2b82e2-a807-49f8-9c77-f12a8b8bde3c-kube-api-access-g94mk\") pod \"8e2b82e2-a807-49f8-9c77-f12a8b8bde3c\" (UID: \"8e2b82e2-a807-49f8-9c77-f12a8b8bde3c\") " Oct 08 14:20:04 crc kubenswrapper[4789]: I1008 14:20:04.958081 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e2b82e2-a807-49f8-9c77-f12a8b8bde3c-kube-api-access-g94mk" (OuterVolumeSpecName: "kube-api-access-g94mk") pod "8e2b82e2-a807-49f8-9c77-f12a8b8bde3c" (UID: "8e2b82e2-a807-49f8-9c77-f12a8b8bde3c"). InnerVolumeSpecName "kube-api-access-g94mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:05 crc kubenswrapper[4789]: I1008 14:20:05.055792 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g94mk\" (UniqueName: \"kubernetes.io/projected/8e2b82e2-a807-49f8-9c77-f12a8b8bde3c-kube-api-access-g94mk\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:05 crc kubenswrapper[4789]: I1008 14:20:05.409314 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4fhfk" event={"ID":"8e2b82e2-a807-49f8-9c77-f12a8b8bde3c","Type":"ContainerDied","Data":"5ba2d8c54d2c30ff5993f40827928ce743434b67c024cc55650a86a6d2005d00"} Oct 08 14:20:05 crc kubenswrapper[4789]: I1008 14:20:05.409354 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ba2d8c54d2c30ff5993f40827928ce743434b67c024cc55650a86a6d2005d00" Oct 08 14:20:05 crc kubenswrapper[4789]: I1008 14:20:05.409472 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4fhfk" Oct 08 14:20:07 crc kubenswrapper[4789]: I1008 14:20:07.518338 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9wtgk" Oct 08 14:20:07 crc kubenswrapper[4789]: I1008 14:20:07.604446 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt28n\" (UniqueName: \"kubernetes.io/projected/19ded6af-86ab-4942-b141-14db479a5986-kube-api-access-jt28n\") pod \"19ded6af-86ab-4942-b141-14db479a5986\" (UID: \"19ded6af-86ab-4942-b141-14db479a5986\") " Oct 08 14:20:07 crc kubenswrapper[4789]: I1008 14:20:07.610868 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ded6af-86ab-4942-b141-14db479a5986-kube-api-access-jt28n" (OuterVolumeSpecName: "kube-api-access-jt28n") pod "19ded6af-86ab-4942-b141-14db479a5986" (UID: "19ded6af-86ab-4942-b141-14db479a5986"). InnerVolumeSpecName "kube-api-access-jt28n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:07 crc kubenswrapper[4789]: I1008 14:20:07.706318 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt28n\" (UniqueName: \"kubernetes.io/projected/19ded6af-86ab-4942-b141-14db479a5986-kube-api-access-jt28n\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:08 crc kubenswrapper[4789]: I1008 14:20:08.439819 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-9wtgk" event={"ID":"19ded6af-86ab-4942-b141-14db479a5986","Type":"ContainerDied","Data":"dc841c4113cdee897526ec57382e95ee7bcdb6e2f9e411201c6c3b39c2358bfb"} Oct 08 14:20:08 crc kubenswrapper[4789]: I1008 14:20:08.440160 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc841c4113cdee897526ec57382e95ee7bcdb6e2f9e411201c6c3b39c2358bfb" Oct 08 14:20:08 crc kubenswrapper[4789]: I1008 14:20:08.439893 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-9wtgk" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.385937 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-e02c-account-create-zrxdw"] Oct 08 14:20:10 crc kubenswrapper[4789]: E1008 14:20:10.386509 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ded6af-86ab-4942-b141-14db479a5986" containerName="mariadb-database-create" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.386523 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ded6af-86ab-4942-b141-14db479a5986" containerName="mariadb-database-create" Oct 08 14:20:10 crc kubenswrapper[4789]: E1008 14:20:10.386538 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f2731e-7573-4376-96bb-e19ac616a706" containerName="mariadb-database-create" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.386544 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f2731e-7573-4376-96bb-e19ac616a706" containerName="mariadb-database-create" Oct 08 14:20:10 crc kubenswrapper[4789]: E1008 14:20:10.386574 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2b82e2-a807-49f8-9c77-f12a8b8bde3c" containerName="mariadb-database-create" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.386580 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2b82e2-a807-49f8-9c77-f12a8b8bde3c" containerName="mariadb-database-create" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.386742 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ded6af-86ab-4942-b141-14db479a5986" containerName="mariadb-database-create" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.386763 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="48f2731e-7573-4376-96bb-e19ac616a706" containerName="mariadb-database-create" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.386779 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e2b82e2-a807-49f8-9c77-f12a8b8bde3c" containerName="mariadb-database-create" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.387322 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e02c-account-create-zrxdw" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.390244 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.392149 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e02c-account-create-zrxdw"] Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.462700 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg7q8\" (UniqueName: \"kubernetes.io/projected/fd52452e-0042-42ae-9c01-649d38f32965-kube-api-access-rg7q8\") pod \"glance-e02c-account-create-zrxdw\" (UID: \"fd52452e-0042-42ae-9c01-649d38f32965\") " pod="openstack/glance-e02c-account-create-zrxdw" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.564249 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg7q8\" (UniqueName: \"kubernetes.io/projected/fd52452e-0042-42ae-9c01-649d38f32965-kube-api-access-rg7q8\") pod \"glance-e02c-account-create-zrxdw\" (UID: \"fd52452e-0042-42ae-9c01-649d38f32965\") " pod="openstack/glance-e02c-account-create-zrxdw" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.582180 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg7q8\" (UniqueName: \"kubernetes.io/projected/fd52452e-0042-42ae-9c01-649d38f32965-kube-api-access-rg7q8\") pod \"glance-e02c-account-create-zrxdw\" (UID: \"fd52452e-0042-42ae-9c01-649d38f32965\") " pod="openstack/glance-e02c-account-create-zrxdw" Oct 08 14:20:10 crc kubenswrapper[4789]: I1008 14:20:10.780908 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e02c-account-create-zrxdw" Oct 08 14:20:11 crc kubenswrapper[4789]: I1008 14:20:11.833830 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-1f09-account-create-4n6fh"] Oct 08 14:20:11 crc kubenswrapper[4789]: I1008 14:20:11.836334 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1f09-account-create-4n6fh" Oct 08 14:20:11 crc kubenswrapper[4789]: I1008 14:20:11.837755 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1f09-account-create-4n6fh"] Oct 08 14:20:11 crc kubenswrapper[4789]: I1008 14:20:11.838421 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 08 14:20:11 crc kubenswrapper[4789]: I1008 14:20:11.886514 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz582\" (UniqueName: \"kubernetes.io/projected/4a6157f7-df72-4d4a-a3fb-b928c430b4ee-kube-api-access-sz582\") pod \"barbican-1f09-account-create-4n6fh\" (UID: \"4a6157f7-df72-4d4a-a3fb-b928c430b4ee\") " pod="openstack/barbican-1f09-account-create-4n6fh" Oct 08 14:20:11 crc kubenswrapper[4789]: I1008 14:20:11.966119 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 08 14:20:11 crc kubenswrapper[4789]: I1008 14:20:11.970742 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 08 14:20:11 crc kubenswrapper[4789]: I1008 14:20:11.988101 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz582\" (UniqueName: \"kubernetes.io/projected/4a6157f7-df72-4d4a-a3fb-b928c430b4ee-kube-api-access-sz582\") pod \"barbican-1f09-account-create-4n6fh\" (UID: \"4a6157f7-df72-4d4a-a3fb-b928c430b4ee\") " pod="openstack/barbican-1f09-account-create-4n6fh" Oct 08 14:20:12 crc kubenswrapper[4789]: I1008 14:20:12.009761 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz582\" (UniqueName: \"kubernetes.io/projected/4a6157f7-df72-4d4a-a3fb-b928c430b4ee-kube-api-access-sz582\") pod \"barbican-1f09-account-create-4n6fh\" (UID: \"4a6157f7-df72-4d4a-a3fb-b928c430b4ee\") " pod="openstack/barbican-1f09-account-create-4n6fh" Oct 08 14:20:12 crc kubenswrapper[4789]: I1008 14:20:12.164210 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1f09-account-create-4n6fh" Oct 08 14:20:12 crc kubenswrapper[4789]: I1008 14:20:12.490845 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 08 14:20:15 crc kubenswrapper[4789]: I1008 14:20:15.778638 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7nqh5" Oct 08 14:20:15 crc kubenswrapper[4789]: I1008 14:20:15.853068 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nt7s\" (UniqueName: \"kubernetes.io/projected/a038fc03-8ce7-4641-a8f1-1e2a748847a9-kube-api-access-4nt7s\") pod \"a038fc03-8ce7-4641-a8f1-1e2a748847a9\" (UID: \"a038fc03-8ce7-4641-a8f1-1e2a748847a9\") " Oct 08 14:20:15 crc kubenswrapper[4789]: I1008 14:20:15.859085 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a038fc03-8ce7-4641-a8f1-1e2a748847a9-kube-api-access-4nt7s" (OuterVolumeSpecName: "kube-api-access-4nt7s") pod "a038fc03-8ce7-4641-a8f1-1e2a748847a9" (UID: "a038fc03-8ce7-4641-a8f1-1e2a748847a9"). InnerVolumeSpecName "kube-api-access-4nt7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:15 crc kubenswrapper[4789]: I1008 14:20:15.954343 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nt7s\" (UniqueName: \"kubernetes.io/projected/a038fc03-8ce7-4641-a8f1-1e2a748847a9-kube-api-access-4nt7s\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:16 crc kubenswrapper[4789]: E1008 14:20:16.371756 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-watcher-api:watcher_latest" Oct 08 14:20:16 crc kubenswrapper[4789]: E1008 14:20:16.371830 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-watcher-api:watcher_latest" Oct 08 14:20:16 crc kubenswrapper[4789]: E1008 14:20:16.372001 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:watcher-db-sync,Image:38.102.83.222:5001/podified-master-centos10/openstack-watcher-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/watcher/watcher.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:watcher-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vb5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-db-sync-6w7kw_openstack(c629d8f6-586d-4f5c-a7ea-2e6912a02b03): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:20:16 crc kubenswrapper[4789]: E1008 14:20:16.373194 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/watcher-db-sync-6w7kw" podUID="c629d8f6-586d-4f5c-a7ea-2e6912a02b03" Oct 08 14:20:16 crc kubenswrapper[4789]: I1008 14:20:16.520904 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7nqh5" Oct 08 14:20:16 crc kubenswrapper[4789]: I1008 14:20:16.521754 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7nqh5" event={"ID":"a038fc03-8ce7-4641-a8f1-1e2a748847a9","Type":"ContainerDied","Data":"32ce0c0e50306a38aa8236b25984940d34f526946de41510443f36af82896825"} Oct 08 14:20:16 crc kubenswrapper[4789]: I1008 14:20:16.522043 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32ce0c0e50306a38aa8236b25984940d34f526946de41510443f36af82896825" Oct 08 14:20:16 crc kubenswrapper[4789]: E1008 14:20:16.525093 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-watcher-api:watcher_latest\\\"\"" pod="openstack/watcher-db-sync-6w7kw" podUID="c629d8f6-586d-4f5c-a7ea-2e6912a02b03" Oct 08 14:20:16 crc kubenswrapper[4789]: I1008 14:20:16.802167 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e02c-account-create-zrxdw"] Oct 08 14:20:16 crc kubenswrapper[4789]: I1008 14:20:16.855077 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1f09-account-create-4n6fh"] Oct 08 14:20:16 crc kubenswrapper[4789]: W1008 14:20:16.857456 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a6157f7_df72_4d4a_a3fb_b928c430b4ee.slice/crio-a53704ca0cc3bfcf57a01dd460d91b4be6a42e0790161f3c575173fd4b380042 WatchSource:0}: Error finding container a53704ca0cc3bfcf57a01dd460d91b4be6a42e0790161f3c575173fd4b380042: Status 404 returned error can't find the container with id a53704ca0cc3bfcf57a01dd460d91b4be6a42e0790161f3c575173fd4b380042 Oct 08 14:20:17 crc kubenswrapper[4789]: I1008 14:20:17.538951 4789 generic.go:334] "Generic (PLEG): container finished" podID="4a6157f7-df72-4d4a-a3fb-b928c430b4ee" containerID="71929a9f9c5be53fcac8c7aaec81ee8fd28dc08fd3116ec63a85ba6d363edeca" exitCode=0 Oct 08 14:20:17 crc kubenswrapper[4789]: I1008 14:20:17.539151 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1f09-account-create-4n6fh" event={"ID":"4a6157f7-df72-4d4a-a3fb-b928c430b4ee","Type":"ContainerDied","Data":"71929a9f9c5be53fcac8c7aaec81ee8fd28dc08fd3116ec63a85ba6d363edeca"} Oct 08 14:20:17 crc kubenswrapper[4789]: I1008 14:20:17.539643 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1f09-account-create-4n6fh" event={"ID":"4a6157f7-df72-4d4a-a3fb-b928c430b4ee","Type":"ContainerStarted","Data":"a53704ca0cc3bfcf57a01dd460d91b4be6a42e0790161f3c575173fd4b380042"} Oct 08 14:20:17 crc kubenswrapper[4789]: I1008 14:20:17.544649 4789 generic.go:334] "Generic (PLEG): container finished" podID="fd52452e-0042-42ae-9c01-649d38f32965" containerID="d99c1ff3eee5ae97cd5a41fd82ae1583be6522fe8c2476158488d5691c0d3b16" exitCode=0 Oct 08 14:20:17 crc kubenswrapper[4789]: I1008 14:20:17.544733 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e02c-account-create-zrxdw" event={"ID":"fd52452e-0042-42ae-9c01-649d38f32965","Type":"ContainerDied","Data":"d99c1ff3eee5ae97cd5a41fd82ae1583be6522fe8c2476158488d5691c0d3b16"} Oct 08 14:20:17 crc kubenswrapper[4789]: I1008 14:20:17.544793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e02c-account-create-zrxdw" event={"ID":"fd52452e-0042-42ae-9c01-649d38f32965","Type":"ContainerStarted","Data":"2f5ef71429aad87526da8d64e84bb544da699920f80c6a689a1663f190c941fd"} Oct 08 14:20:17 crc kubenswrapper[4789]: I1008 14:20:17.546569 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-srh88" event={"ID":"187c395c-7b8d-45e7-b96c-a93cea38b025","Type":"ContainerStarted","Data":"fd24d64491398970d54cdd13f113442754d73038f8bab41f05c197728fe01471"} Oct 08 14:20:17 crc kubenswrapper[4789]: I1008 14:20:17.608702 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-srh88" podStartSLOduration=2.5812601280000003 podStartE2EDuration="15.608683735s" podCreationTimestamp="2025-10-08 14:20:02 +0000 UTC" firstStartedPulling="2025-10-08 14:20:03.335060615 +0000 UTC m=+1143.241808097" lastFinishedPulling="2025-10-08 14:20:16.362484212 +0000 UTC m=+1156.269231704" observedRunningTime="2025-10-08 14:20:17.600575896 +0000 UTC m=+1157.507323398" watchObservedRunningTime="2025-10-08 14:20:17.608683735 +0000 UTC m=+1157.515431227" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.012598 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e02c-account-create-zrxdw" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.017616 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1f09-account-create-4n6fh" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.103467 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg7q8\" (UniqueName: \"kubernetes.io/projected/fd52452e-0042-42ae-9c01-649d38f32965-kube-api-access-rg7q8\") pod \"fd52452e-0042-42ae-9c01-649d38f32965\" (UID: \"fd52452e-0042-42ae-9c01-649d38f32965\") " Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.103592 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz582\" (UniqueName: \"kubernetes.io/projected/4a6157f7-df72-4d4a-a3fb-b928c430b4ee-kube-api-access-sz582\") pod \"4a6157f7-df72-4d4a-a3fb-b928c430b4ee\" (UID: \"4a6157f7-df72-4d4a-a3fb-b928c430b4ee\") " Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.117806 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd52452e-0042-42ae-9c01-649d38f32965-kube-api-access-rg7q8" (OuterVolumeSpecName: "kube-api-access-rg7q8") pod "fd52452e-0042-42ae-9c01-649d38f32965" (UID: "fd52452e-0042-42ae-9c01-649d38f32965"). InnerVolumeSpecName "kube-api-access-rg7q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.120887 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a6157f7-df72-4d4a-a3fb-b928c430b4ee-kube-api-access-sz582" (OuterVolumeSpecName: "kube-api-access-sz582") pod "4a6157f7-df72-4d4a-a3fb-b928c430b4ee" (UID: "4a6157f7-df72-4d4a-a3fb-b928c430b4ee"). InnerVolumeSpecName "kube-api-access-sz582". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.205826 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg7q8\" (UniqueName: \"kubernetes.io/projected/fd52452e-0042-42ae-9c01-649d38f32965-kube-api-access-rg7q8\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.205870 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz582\" (UniqueName: \"kubernetes.io/projected/4a6157f7-df72-4d4a-a3fb-b928c430b4ee-kube-api-access-sz582\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.564824 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1f09-account-create-4n6fh" event={"ID":"4a6157f7-df72-4d4a-a3fb-b928c430b4ee","Type":"ContainerDied","Data":"a53704ca0cc3bfcf57a01dd460d91b4be6a42e0790161f3c575173fd4b380042"} Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.564867 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1f09-account-create-4n6fh" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.564875 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a53704ca0cc3bfcf57a01dd460d91b4be6a42e0790161f3c575173fd4b380042" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.566673 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e02c-account-create-zrxdw" event={"ID":"fd52452e-0042-42ae-9c01-649d38f32965","Type":"ContainerDied","Data":"2f5ef71429aad87526da8d64e84bb544da699920f80c6a689a1663f190c941fd"} Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.566734 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f5ef71429aad87526da8d64e84bb544da699920f80c6a689a1663f190c941fd" Oct 08 14:20:19 crc kubenswrapper[4789]: I1008 14:20:19.566743 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e02c-account-create-zrxdw" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.446161 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-5c9mc"] Oct 08 14:20:20 crc kubenswrapper[4789]: E1008 14:20:20.446817 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a6157f7-df72-4d4a-a3fb-b928c430b4ee" containerName="mariadb-account-create" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.446831 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a6157f7-df72-4d4a-a3fb-b928c430b4ee" containerName="mariadb-account-create" Oct 08 14:20:20 crc kubenswrapper[4789]: E1008 14:20:20.446854 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd52452e-0042-42ae-9c01-649d38f32965" containerName="mariadb-account-create" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.446861 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd52452e-0042-42ae-9c01-649d38f32965" containerName="mariadb-account-create" Oct 08 14:20:20 crc kubenswrapper[4789]: E1008 14:20:20.446888 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a038fc03-8ce7-4641-a8f1-1e2a748847a9" containerName="mariadb-database-create" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.446896 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a038fc03-8ce7-4641-a8f1-1e2a748847a9" containerName="mariadb-database-create" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.447071 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a6157f7-df72-4d4a-a3fb-b928c430b4ee" containerName="mariadb-account-create" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.447083 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a038fc03-8ce7-4641-a8f1-1e2a748847a9" containerName="mariadb-database-create" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.447102 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd52452e-0042-42ae-9c01-649d38f32965" containerName="mariadb-account-create" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.447699 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.450024 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.450066 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rckbp" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.460601 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5c9mc"] Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.525870 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-config-data\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.525929 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-db-sync-config-data\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.526058 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-combined-ca-bundle\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.526088 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txzxz\" (UniqueName: \"kubernetes.io/projected/5d62cf1d-f82b-4eed-8500-2614367a4e3d-kube-api-access-txzxz\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.628249 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-db-sync-config-data\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.628342 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-combined-ca-bundle\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.628374 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txzxz\" (UniqueName: \"kubernetes.io/projected/5d62cf1d-f82b-4eed-8500-2614367a4e3d-kube-api-access-txzxz\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.628499 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-config-data\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.632170 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-db-sync-config-data\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.635598 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-config-data\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.636745 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-combined-ca-bundle\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.646115 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txzxz\" (UniqueName: \"kubernetes.io/projected/5d62cf1d-f82b-4eed-8500-2614367a4e3d-kube-api-access-txzxz\") pod \"glance-db-sync-5c9mc\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:20 crc kubenswrapper[4789]: I1008 14:20:20.769947 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5c9mc" Oct 08 14:20:21 crc kubenswrapper[4789]: I1008 14:20:21.344756 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5c9mc"] Oct 08 14:20:21 crc kubenswrapper[4789]: I1008 14:20:21.583574 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5c9mc" event={"ID":"5d62cf1d-f82b-4eed-8500-2614367a4e3d","Type":"ContainerStarted","Data":"19b87444e11c1e7411fdfde5ceef1e8091450a6c9dc2973af2e2b4cb59352ca2"} Oct 08 14:20:21 crc kubenswrapper[4789]: I1008 14:20:21.964440 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f16b-account-create-vkgg2"] Oct 08 14:20:21 crc kubenswrapper[4789]: I1008 14:20:21.965613 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f16b-account-create-vkgg2" Oct 08 14:20:21 crc kubenswrapper[4789]: I1008 14:20:21.971059 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 08 14:20:21 crc kubenswrapper[4789]: I1008 14:20:21.987622 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f16b-account-create-vkgg2"] Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.156848 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hmb4\" (UniqueName: \"kubernetes.io/projected/52156123-e5c0-4e28-b9dc-f972398de562-kube-api-access-4hmb4\") pod \"cinder-f16b-account-create-vkgg2\" (UID: \"52156123-e5c0-4e28-b9dc-f972398de562\") " pod="openstack/cinder-f16b-account-create-vkgg2" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.157363 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-266d-account-create-tc8nw"] Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.158467 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-266d-account-create-tc8nw" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.160673 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.171766 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-266d-account-create-tc8nw"] Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.258819 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zrsh\" (UniqueName: \"kubernetes.io/projected/1a7857f4-1cb7-4aed-b42f-401f54afb6c8-kube-api-access-8zrsh\") pod \"neutron-266d-account-create-tc8nw\" (UID: \"1a7857f4-1cb7-4aed-b42f-401f54afb6c8\") " pod="openstack/neutron-266d-account-create-tc8nw" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.258901 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hmb4\" (UniqueName: \"kubernetes.io/projected/52156123-e5c0-4e28-b9dc-f972398de562-kube-api-access-4hmb4\") pod \"cinder-f16b-account-create-vkgg2\" (UID: \"52156123-e5c0-4e28-b9dc-f972398de562\") " pod="openstack/cinder-f16b-account-create-vkgg2" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.278491 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hmb4\" (UniqueName: \"kubernetes.io/projected/52156123-e5c0-4e28-b9dc-f972398de562-kube-api-access-4hmb4\") pod \"cinder-f16b-account-create-vkgg2\" (UID: \"52156123-e5c0-4e28-b9dc-f972398de562\") " pod="openstack/cinder-f16b-account-create-vkgg2" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.308939 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f16b-account-create-vkgg2" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.359975 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zrsh\" (UniqueName: \"kubernetes.io/projected/1a7857f4-1cb7-4aed-b42f-401f54afb6c8-kube-api-access-8zrsh\") pod \"neutron-266d-account-create-tc8nw\" (UID: \"1a7857f4-1cb7-4aed-b42f-401f54afb6c8\") " pod="openstack/neutron-266d-account-create-tc8nw" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.381443 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zrsh\" (UniqueName: \"kubernetes.io/projected/1a7857f4-1cb7-4aed-b42f-401f54afb6c8-kube-api-access-8zrsh\") pod \"neutron-266d-account-create-tc8nw\" (UID: \"1a7857f4-1cb7-4aed-b42f-401f54afb6c8\") " pod="openstack/neutron-266d-account-create-tc8nw" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.488461 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-266d-account-create-tc8nw" Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.722727 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f16b-account-create-vkgg2"] Oct 08 14:20:22 crc kubenswrapper[4789]: I1008 14:20:22.889837 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-266d-account-create-tc8nw"] Oct 08 14:20:22 crc kubenswrapper[4789]: W1008 14:20:22.900440 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a7857f4_1cb7_4aed_b42f_401f54afb6c8.slice/crio-eaf73e9349b7150dc94b7a2fc5e7dd81e9961803f71d25197d7d4e44be5573c0 WatchSource:0}: Error finding container eaf73e9349b7150dc94b7a2fc5e7dd81e9961803f71d25197d7d4e44be5573c0: Status 404 returned error can't find the container with id eaf73e9349b7150dc94b7a2fc5e7dd81e9961803f71d25197d7d4e44be5573c0 Oct 08 14:20:23 crc kubenswrapper[4789]: I1008 14:20:23.601113 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f16b-account-create-vkgg2" event={"ID":"52156123-e5c0-4e28-b9dc-f972398de562","Type":"ContainerStarted","Data":"76a8ccf7019532e6e1c615e7262b05ef49e6789012cf931985df04ad37dea96a"} Oct 08 14:20:23 crc kubenswrapper[4789]: I1008 14:20:23.601175 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f16b-account-create-vkgg2" event={"ID":"52156123-e5c0-4e28-b9dc-f972398de562","Type":"ContainerStarted","Data":"5cfab7769807c8b8f18117b365c458118e99ef3b764046a813ab211df9e54d66"} Oct 08 14:20:23 crc kubenswrapper[4789]: I1008 14:20:23.603265 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-266d-account-create-tc8nw" event={"ID":"1a7857f4-1cb7-4aed-b42f-401f54afb6c8","Type":"ContainerStarted","Data":"6c56f8c678f13de06c504d4b89c16ed3c33342a592c3dd2df0c28de159d97fc7"} Oct 08 14:20:23 crc kubenswrapper[4789]: I1008 14:20:23.603345 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-266d-account-create-tc8nw" event={"ID":"1a7857f4-1cb7-4aed-b42f-401f54afb6c8","Type":"ContainerStarted","Data":"eaf73e9349b7150dc94b7a2fc5e7dd81e9961803f71d25197d7d4e44be5573c0"} Oct 08 14:20:23 crc kubenswrapper[4789]: I1008 14:20:23.618722 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-f16b-account-create-vkgg2" podStartSLOduration=2.618703579 podStartE2EDuration="2.618703579s" podCreationTimestamp="2025-10-08 14:20:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:20:23.615042289 +0000 UTC m=+1163.521789781" watchObservedRunningTime="2025-10-08 14:20:23.618703579 +0000 UTC m=+1163.525451071" Oct 08 14:20:23 crc kubenswrapper[4789]: I1008 14:20:23.629661 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-266d-account-create-tc8nw" podStartSLOduration=1.629640455 podStartE2EDuration="1.629640455s" podCreationTimestamp="2025-10-08 14:20:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:20:23.62797034 +0000 UTC m=+1163.534717832" watchObservedRunningTime="2025-10-08 14:20:23.629640455 +0000 UTC m=+1163.536387947" Oct 08 14:20:24 crc kubenswrapper[4789]: I1008 14:20:24.612367 4789 generic.go:334] "Generic (PLEG): container finished" podID="52156123-e5c0-4e28-b9dc-f972398de562" containerID="76a8ccf7019532e6e1c615e7262b05ef49e6789012cf931985df04ad37dea96a" exitCode=0 Oct 08 14:20:24 crc kubenswrapper[4789]: I1008 14:20:24.612452 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f16b-account-create-vkgg2" event={"ID":"52156123-e5c0-4e28-b9dc-f972398de562","Type":"ContainerDied","Data":"76a8ccf7019532e6e1c615e7262b05ef49e6789012cf931985df04ad37dea96a"} Oct 08 14:20:24 crc kubenswrapper[4789]: I1008 14:20:24.614140 4789 generic.go:334] "Generic (PLEG): container finished" podID="1a7857f4-1cb7-4aed-b42f-401f54afb6c8" containerID="6c56f8c678f13de06c504d4b89c16ed3c33342a592c3dd2df0c28de159d97fc7" exitCode=0 Oct 08 14:20:24 crc kubenswrapper[4789]: I1008 14:20:24.614171 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-266d-account-create-tc8nw" event={"ID":"1a7857f4-1cb7-4aed-b42f-401f54afb6c8","Type":"ContainerDied","Data":"6c56f8c678f13de06c504d4b89c16ed3c33342a592c3dd2df0c28de159d97fc7"} Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.006804 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-266d-account-create-tc8nw" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.098891 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f16b-account-create-vkgg2" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.130268 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zrsh\" (UniqueName: \"kubernetes.io/projected/1a7857f4-1cb7-4aed-b42f-401f54afb6c8-kube-api-access-8zrsh\") pod \"1a7857f4-1cb7-4aed-b42f-401f54afb6c8\" (UID: \"1a7857f4-1cb7-4aed-b42f-401f54afb6c8\") " Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.151184 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a7857f4-1cb7-4aed-b42f-401f54afb6c8-kube-api-access-8zrsh" (OuterVolumeSpecName: "kube-api-access-8zrsh") pod "1a7857f4-1cb7-4aed-b42f-401f54afb6c8" (UID: "1a7857f4-1cb7-4aed-b42f-401f54afb6c8"). InnerVolumeSpecName "kube-api-access-8zrsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.232469 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hmb4\" (UniqueName: \"kubernetes.io/projected/52156123-e5c0-4e28-b9dc-f972398de562-kube-api-access-4hmb4\") pod \"52156123-e5c0-4e28-b9dc-f972398de562\" (UID: \"52156123-e5c0-4e28-b9dc-f972398de562\") " Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.233527 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zrsh\" (UniqueName: \"kubernetes.io/projected/1a7857f4-1cb7-4aed-b42f-401f54afb6c8-kube-api-access-8zrsh\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.236672 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52156123-e5c0-4e28-b9dc-f972398de562-kube-api-access-4hmb4" (OuterVolumeSpecName: "kube-api-access-4hmb4") pod "52156123-e5c0-4e28-b9dc-f972398de562" (UID: "52156123-e5c0-4e28-b9dc-f972398de562"). InnerVolumeSpecName "kube-api-access-4hmb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.334694 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hmb4\" (UniqueName: \"kubernetes.io/projected/52156123-e5c0-4e28-b9dc-f972398de562-kube-api-access-4hmb4\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.434209 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.434270 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.434313 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.434959 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8550892ad865dfdb5faa70bb82d5784fbdac5d69789616846b92c98439e8b06"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.435048 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://b8550892ad865dfdb5faa70bb82d5784fbdac5d69789616846b92c98439e8b06" gracePeriod=600 Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.631230 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f16b-account-create-vkgg2" event={"ID":"52156123-e5c0-4e28-b9dc-f972398de562","Type":"ContainerDied","Data":"5cfab7769807c8b8f18117b365c458118e99ef3b764046a813ab211df9e54d66"} Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.631266 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cfab7769807c8b8f18117b365c458118e99ef3b764046a813ab211df9e54d66" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.631318 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f16b-account-create-vkgg2" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.648372 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-266d-account-create-tc8nw" event={"ID":"1a7857f4-1cb7-4aed-b42f-401f54afb6c8","Type":"ContainerDied","Data":"eaf73e9349b7150dc94b7a2fc5e7dd81e9961803f71d25197d7d4e44be5573c0"} Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.648416 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaf73e9349b7150dc94b7a2fc5e7dd81e9961803f71d25197d7d4e44be5573c0" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.648492 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-266d-account-create-tc8nw" Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.653240 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="b8550892ad865dfdb5faa70bb82d5784fbdac5d69789616846b92c98439e8b06" exitCode=0 Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.653552 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"b8550892ad865dfdb5faa70bb82d5784fbdac5d69789616846b92c98439e8b06"} Oct 08 14:20:26 crc kubenswrapper[4789]: I1008 14:20:26.653590 4789 scope.go:117] "RemoveContainer" containerID="49beaccc4d4e056fc391f28224570c17529721b7b590ae8c4d56a12f0b6f40ae" Oct 08 14:20:30 crc kubenswrapper[4789]: I1008 14:20:30.688571 4789 generic.go:334] "Generic (PLEG): container finished" podID="187c395c-7b8d-45e7-b96c-a93cea38b025" containerID="fd24d64491398970d54cdd13f113442754d73038f8bab41f05c197728fe01471" exitCode=0 Oct 08 14:20:30 crc kubenswrapper[4789]: I1008 14:20:30.688665 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-srh88" event={"ID":"187c395c-7b8d-45e7-b96c-a93cea38b025","Type":"ContainerDied","Data":"fd24d64491398970d54cdd13f113442754d73038f8bab41f05c197728fe01471"} Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.440887 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:38 crc kubenswrapper[4789]: E1008 14:20:38.561544 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Oct 08 14:20:38 crc kubenswrapper[4789]: E1008 14:20:38.561621 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Oct 08 14:20:38 crc kubenswrapper[4789]: E1008 14:20:38.561825 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:38.102.83.222:5001/podified-master-centos10/openstack-glance-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-txzxz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-5c9mc_openstack(5d62cf1d-f82b-4eed-8500-2614367a4e3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:20:38 crc kubenswrapper[4789]: E1008 14:20:38.563082 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-5c9mc" podUID="5d62cf1d-f82b-4eed-8500-2614367a4e3d" Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.580846 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-486gf\" (UniqueName: \"kubernetes.io/projected/187c395c-7b8d-45e7-b96c-a93cea38b025-kube-api-access-486gf\") pod \"187c395c-7b8d-45e7-b96c-a93cea38b025\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.580909 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-combined-ca-bundle\") pod \"187c395c-7b8d-45e7-b96c-a93cea38b025\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.581187 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-config-data\") pod \"187c395c-7b8d-45e7-b96c-a93cea38b025\" (UID: \"187c395c-7b8d-45e7-b96c-a93cea38b025\") " Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.586184 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/187c395c-7b8d-45e7-b96c-a93cea38b025-kube-api-access-486gf" (OuterVolumeSpecName: "kube-api-access-486gf") pod "187c395c-7b8d-45e7-b96c-a93cea38b025" (UID: "187c395c-7b8d-45e7-b96c-a93cea38b025"). InnerVolumeSpecName "kube-api-access-486gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.608610 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "187c395c-7b8d-45e7-b96c-a93cea38b025" (UID: "187c395c-7b8d-45e7-b96c-a93cea38b025"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.623187 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-config-data" (OuterVolumeSpecName: "config-data") pod "187c395c-7b8d-45e7-b96c-a93cea38b025" (UID: "187c395c-7b8d-45e7-b96c-a93cea38b025"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.682910 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.682948 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-486gf\" (UniqueName: \"kubernetes.io/projected/187c395c-7b8d-45e7-b96c-a93cea38b025-kube-api-access-486gf\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.682958 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/187c395c-7b8d-45e7-b96c-a93cea38b025-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.762831 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"26322f12c01ca82f351a4ae80e71651fb887b0762ae10b5b893ab0870c97be77"} Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.764964 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-srh88" Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.765105 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-srh88" event={"ID":"187c395c-7b8d-45e7-b96c-a93cea38b025","Type":"ContainerDied","Data":"b7d97602be651ccee34c51aeb9f5a8e445064a73c4596a2ce4ec5718f2174b47"} Oct 08 14:20:38 crc kubenswrapper[4789]: I1008 14:20:38.765190 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7d97602be651ccee34c51aeb9f5a8e445064a73c4596a2ce4ec5718f2174b47" Oct 08 14:20:38 crc kubenswrapper[4789]: E1008 14:20:38.984285 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-glance-api:watcher_latest\\\"\"" pod="openstack/glance-db-sync-5c9mc" podUID="5d62cf1d-f82b-4eed-8500-2614367a4e3d" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.716909 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b7bd99c95-g2wcj"] Oct 08 14:20:39 crc kubenswrapper[4789]: E1008 14:20:39.717688 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a7857f4-1cb7-4aed-b42f-401f54afb6c8" containerName="mariadb-account-create" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.717709 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a7857f4-1cb7-4aed-b42f-401f54afb6c8" containerName="mariadb-account-create" Oct 08 14:20:39 crc kubenswrapper[4789]: E1008 14:20:39.717771 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="187c395c-7b8d-45e7-b96c-a93cea38b025" containerName="keystone-db-sync" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.717781 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="187c395c-7b8d-45e7-b96c-a93cea38b025" containerName="keystone-db-sync" Oct 08 14:20:39 crc kubenswrapper[4789]: E1008 14:20:39.717800 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52156123-e5c0-4e28-b9dc-f972398de562" containerName="mariadb-account-create" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.717807 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="52156123-e5c0-4e28-b9dc-f972398de562" containerName="mariadb-account-create" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.718017 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a7857f4-1cb7-4aed-b42f-401f54afb6c8" containerName="mariadb-account-create" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.718039 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="187c395c-7b8d-45e7-b96c-a93cea38b025" containerName="keystone-db-sync" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.718058 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="52156123-e5c0-4e28-b9dc-f972398de562" containerName="mariadb-account-create" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.719171 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.785403 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fdzqc"] Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.801037 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.799627 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-6w7kw" event={"ID":"c629d8f6-586d-4f5c-a7ea-2e6912a02b03","Type":"ContainerStarted","Data":"54d9d255673843e618360f4e57b1dfe2bfe4e54a8bf1e14f224ed47b041609cc"} Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.805658 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8h2vv" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.805954 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.806070 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.806088 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.817183 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b7bd99c95-g2wcj"] Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.839063 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fdzqc"] Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.875127 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-6w7kw" podStartSLOduration=2.030733469 podStartE2EDuration="37.875105997s" podCreationTimestamp="2025-10-08 14:20:02 +0000 UTC" firstStartedPulling="2025-10-08 14:20:03.141506879 +0000 UTC m=+1143.048254371" lastFinishedPulling="2025-10-08 14:20:38.985879407 +0000 UTC m=+1178.892626899" observedRunningTime="2025-10-08 14:20:39.866476753 +0000 UTC m=+1179.773224245" watchObservedRunningTime="2025-10-08 14:20:39.875105997 +0000 UTC m=+1179.781853489" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.916849 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-scripts\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.916920 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-config-data\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.916975 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq5hh\" (UniqueName: \"kubernetes.io/projected/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-kube-api-access-lq5hh\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.917047 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-svc\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.917087 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-swift-storage-0\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.917111 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-config\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.917130 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-sb\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.917152 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-nb\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.917211 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-combined-ca-bundle\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.917238 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9wdd\" (UniqueName: \"kubernetes.io/projected/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-kube-api-access-t9wdd\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.917254 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-fernet-keys\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.917300 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-credential-keys\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.924188 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-p54dp"] Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.925357 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.934621 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xnlqp" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.937319 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-p54dp"] Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.939754 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.939977 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.966176 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-597764f7c9-2zlgs"] Oct 08 14:20:39 crc kubenswrapper[4789]: I1008 14:20:39.980414 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.005299 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.005538 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.005674 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-v69wf" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020043 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-credential-keys\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020115 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-scripts\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020161 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-config-data\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020206 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq5hh\" (UniqueName: \"kubernetes.io/projected/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-kube-api-access-lq5hh\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020242 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-svc\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020309 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-swift-storage-0\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020352 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-config\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020380 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-sb\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020417 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-nb\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020468 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-combined-ca-bundle\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020500 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9wdd\" (UniqueName: \"kubernetes.io/projected/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-kube-api-access-t9wdd\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.020524 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-fernet-keys\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.021849 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-sb\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.028005 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-svc\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.031815 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-fernet-keys\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.032132 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-credential-keys\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.033062 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.033370 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-config-data\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.033896 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-combined-ca-bundle\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.034651 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-swift-storage-0\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.035137 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-scripts\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.035549 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-config\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.035753 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-nb\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.069831 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-597764f7c9-2zlgs"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.080859 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq5hh\" (UniqueName: \"kubernetes.io/projected/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-kube-api-access-lq5hh\") pod \"keystone-bootstrap-fdzqc\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.089607 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9wdd\" (UniqueName: \"kubernetes.io/projected/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-kube-api-access-t9wdd\") pod \"dnsmasq-dns-b7bd99c95-g2wcj\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.124879 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-scripts\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.125156 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-combined-ca-bundle\") pod \"neutron-db-sync-p54dp\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.125243 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-config\") pod \"neutron-db-sync-p54dp\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.125319 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5txnt\" (UniqueName: \"kubernetes.io/projected/0b096b2c-0d8d-44bc-9275-e712b6e622b8-kube-api-access-5txnt\") pod \"neutron-db-sync-p54dp\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.125420 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6m4p\" (UniqueName: \"kubernetes.io/projected/c398271d-3875-4a64-bcf3-6427e8a8762f-kube-api-access-f6m4p\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.125490 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-config-data\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.125583 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c398271d-3875-4a64-bcf3-6427e8a8762f-logs\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.125689 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c398271d-3875-4a64-bcf3-6427e8a8762f-horizon-secret-key\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.147159 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.161120 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-tblv4"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.168455 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.175091 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.175523 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-qw9dn" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.175908 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.217534 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tblv4"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230569 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-scripts\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230641 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-etc-machine-id\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230672 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-scripts\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230714 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-config-data\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230781 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-combined-ca-bundle\") pod \"neutron-db-sync-p54dp\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230812 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-combined-ca-bundle\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230846 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-config\") pod \"neutron-db-sync-p54dp\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230871 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5txnt\" (UniqueName: \"kubernetes.io/projected/0b096b2c-0d8d-44bc-9275-e712b6e622b8-kube-api-access-5txnt\") pod \"neutron-db-sync-p54dp\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230894 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs6qq\" (UniqueName: \"kubernetes.io/projected/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-kube-api-access-bs6qq\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230939 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6m4p\" (UniqueName: \"kubernetes.io/projected/c398271d-3875-4a64-bcf3-6427e8a8762f-kube-api-access-f6m4p\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.230957 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-config-data\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.231044 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c398271d-3875-4a64-bcf3-6427e8a8762f-logs\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.231121 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-db-sync-config-data\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.231149 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c398271d-3875-4a64-bcf3-6427e8a8762f-horizon-secret-key\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.232544 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-scripts\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.234929 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c398271d-3875-4a64-bcf3-6427e8a8762f-logs\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.239456 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c398271d-3875-4a64-bcf3-6427e8a8762f-horizon-secret-key\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.241906 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-hks4r"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.244469 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.250320 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-config-data\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.250399 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hks4r"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.251348 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-combined-ca-bundle\") pod \"neutron-db-sync-p54dp\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.256544 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.256726 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-config\") pod \"neutron-db-sync-p54dp\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.256763 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-svstf" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.280048 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6m4p\" (UniqueName: \"kubernetes.io/projected/c398271d-3875-4a64-bcf3-6427e8a8762f-kube-api-access-f6m4p\") pod \"horizon-597764f7c9-2zlgs\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.280133 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5txnt\" (UniqueName: \"kubernetes.io/projected/0b096b2c-0d8d-44bc-9275-e712b6e622b8-kube-api-access-5txnt\") pod \"neutron-db-sync-p54dp\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.321222 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.334093 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-db-sync-config-data\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.334145 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-scripts\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.334179 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-etc-machine-id\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.334206 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-config-data\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.334238 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-combined-ca-bundle\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.334269 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs6qq\" (UniqueName: \"kubernetes.io/projected/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-kube-api-access-bs6qq\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.334868 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-etc-machine-id\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.344460 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-db-sync-config-data\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.351950 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-combined-ca-bundle\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.351952 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-config-data\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.353257 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-scripts\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.386201 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b7bd99c95-g2wcj"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.387144 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.391773 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs6qq\" (UniqueName: \"kubernetes.io/projected/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-kube-api-access-bs6qq\") pod \"cinder-db-sync-tblv4\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.418788 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-flz9d"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.419971 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.423672 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.423881 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.424004 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-f8hg9" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.442150 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-config-data\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.442191 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-combined-ca-bundle\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.442214 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-scripts\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.442255 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hzzj\" (UniqueName: \"kubernetes.io/projected/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-kube-api-access-4hzzj\") pod \"barbican-db-sync-hks4r\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.442296 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-db-sync-config-data\") pod \"barbican-db-sync-hks4r\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.453920 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-logs\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.454160 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-combined-ca-bundle\") pod \"barbican-db-sync-hks4r\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.454234 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkf94\" (UniqueName: \"kubernetes.io/projected/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-kube-api-access-nkf94\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.459342 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.461512 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.465524 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.465747 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.481090 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-flz9d"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.500672 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5d7cb85455-q84nd"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.519706 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.541892 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tblv4" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.551840 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558716 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-combined-ca-bundle\") pod \"barbican-db-sync-hks4r\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558771 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bswhx\" (UniqueName: \"kubernetes.io/projected/0410f74c-c9df-4218-9d63-790a093c1213-kube-api-access-bswhx\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558808 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkf94\" (UniqueName: \"kubernetes.io/projected/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-kube-api-access-nkf94\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558833 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-config-data\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558855 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-combined-ca-bundle\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558871 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-scripts\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558893 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-config-data\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558918 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558939 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hzzj\" (UniqueName: \"kubernetes.io/projected/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-kube-api-access-4hzzj\") pod \"barbican-db-sync-hks4r\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.558972 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-config-data\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559012 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0410f74c-c9df-4218-9d63-790a093c1213-horizon-secret-key\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559031 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-scripts\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559048 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-db-sync-config-data\") pod \"barbican-db-sync-hks4r\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559072 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2czbg\" (UniqueName: \"kubernetes.io/projected/40b9fe83-3af2-413e-b7fd-cabe53206482-kube-api-access-2czbg\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559107 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-logs\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559123 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0410f74c-c9df-4218-9d63-790a093c1213-logs\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559141 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-run-httpd\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559170 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-scripts\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559193 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-log-httpd\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.559229 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.561601 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-logs\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.567965 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p54dp" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.569809 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-scripts\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.580046 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-db-sync-config-data\") pod \"barbican-db-sync-hks4r\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.580514 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-combined-ca-bundle\") pod \"barbican-db-sync-hks4r\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.590356 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d44448797-vjpbw"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.591764 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.593585 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkf94\" (UniqueName: \"kubernetes.io/projected/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-kube-api-access-nkf94\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.593632 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hzzj\" (UniqueName: \"kubernetes.io/projected/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-kube-api-access-4hzzj\") pod \"barbican-db-sync-hks4r\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.594128 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-combined-ca-bundle\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.594146 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-config-data\") pod \"placement-db-sync-flz9d\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.605524 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hks4r" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.607419 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d7cb85455-q84nd"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.650798 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-flz9d" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.657746 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d44448797-vjpbw"] Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661617 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-config-data\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661647 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0410f74c-c9df-4218-9d63-790a093c1213-horizon-secret-key\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661671 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-sb\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661691 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-scripts\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661707 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-config\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661733 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2czbg\" (UniqueName: \"kubernetes.io/projected/40b9fe83-3af2-413e-b7fd-cabe53206482-kube-api-access-2czbg\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661760 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0410f74c-c9df-4218-9d63-790a093c1213-logs\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661780 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-run-httpd\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661809 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-scripts\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661830 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-log-httpd\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661844 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-swift-storage-0\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661864 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-nb\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661878 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-svc\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661905 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661927 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bswhx\" (UniqueName: \"kubernetes.io/projected/0410f74c-c9df-4218-9d63-790a093c1213-kube-api-access-bswhx\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.661970 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-config-data\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.662006 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fnlx\" (UniqueName: \"kubernetes.io/projected/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-kube-api-access-8fnlx\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.662027 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.663169 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-log-httpd\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.663646 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-scripts\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.665666 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-config-data\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.668237 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.668537 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.670299 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-config-data\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.679495 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0410f74c-c9df-4218-9d63-790a093c1213-logs\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.682408 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-run-httpd\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.690533 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0410f74c-c9df-4218-9d63-790a093c1213-horizon-secret-key\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.693947 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-scripts\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.703239 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2czbg\" (UniqueName: \"kubernetes.io/projected/40b9fe83-3af2-413e-b7fd-cabe53206482-kube-api-access-2czbg\") pod \"ceilometer-0\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " pod="openstack/ceilometer-0" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.704099 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bswhx\" (UniqueName: \"kubernetes.io/projected/0410f74c-c9df-4218-9d63-790a093c1213-kube-api-access-bswhx\") pod \"horizon-5d7cb85455-q84nd\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.765345 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-swift-storage-0\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.768034 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-nb\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.771789 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-swift-storage-0\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.765424 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-nb\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.771911 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-svc\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.772586 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fnlx\" (UniqueName: \"kubernetes.io/projected/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-kube-api-access-8fnlx\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.772817 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-sb\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.772922 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-config\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.776268 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-svc\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.778126 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-config\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.781835 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-sb\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.795746 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fnlx\" (UniqueName: \"kubernetes.io/projected/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-kube-api-access-8fnlx\") pod \"dnsmasq-dns-6d44448797-vjpbw\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:40 crc kubenswrapper[4789]: I1008 14:20:40.971241 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.001782 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fdzqc"] Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.002502 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.022885 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.140117 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-597764f7c9-2zlgs"] Oct 08 14:20:41 crc kubenswrapper[4789]: W1008 14:20:41.172292 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc398271d_3875_4a64_bcf3_6427e8a8762f.slice/crio-b6b5dddf8613b5069152e755eae503da6d84967f52045a1f140e1d7e2677aa9b WatchSource:0}: Error finding container b6b5dddf8613b5069152e755eae503da6d84967f52045a1f140e1d7e2677aa9b: Status 404 returned error can't find the container with id b6b5dddf8613b5069152e755eae503da6d84967f52045a1f140e1d7e2677aa9b Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.358510 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b7bd99c95-g2wcj"] Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.656205 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-flz9d"] Oct 08 14:20:41 crc kubenswrapper[4789]: W1008 14:20:41.665440 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac7cd0ec_3f8d_4afa_84de_cfbb47a7fab9.slice/crio-92e51d8741296d37282ec649137fa18b40ffd6eee17458c0bca3394cdad5a180 WatchSource:0}: Error finding container 92e51d8741296d37282ec649137fa18b40ffd6eee17458c0bca3394cdad5a180: Status 404 returned error can't find the container with id 92e51d8741296d37282ec649137fa18b40ffd6eee17458c0bca3394cdad5a180 Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.689784 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-p54dp"] Oct 08 14:20:41 crc kubenswrapper[4789]: W1008 14:20:41.691861 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b096b2c_0d8d_44bc_9275_e712b6e622b8.slice/crio-10fda8566c16914cc6daa7620204ac1ae3b0a09b92dd436053da2afadbf4e3de WatchSource:0}: Error finding container 10fda8566c16914cc6daa7620204ac1ae3b0a09b92dd436053da2afadbf4e3de: Status 404 returned error can't find the container with id 10fda8566c16914cc6daa7620204ac1ae3b0a09b92dd436053da2afadbf4e3de Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.721302 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tblv4"] Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.739156 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hks4r"] Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.850700 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" event={"ID":"9febac0c-7e85-4ee7-a5e8-6c65de13bda7","Type":"ContainerStarted","Data":"cdff2e70266568360f01c3538531ee0880f1f43fdc572cee6aba3a93c555ca9d"} Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.850769 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" event={"ID":"9febac0c-7e85-4ee7-a5e8-6c65de13bda7","Type":"ContainerStarted","Data":"452efa32bb5048a47bbbf20e298e82ca08bea4fd02817131674f92c68597d4cd"} Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.850939 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" podUID="9febac0c-7e85-4ee7-a5e8-6c65de13bda7" containerName="init" containerID="cri-o://cdff2e70266568360f01c3538531ee0880f1f43fdc572cee6aba3a93c555ca9d" gracePeriod=10 Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.856838 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fdzqc" event={"ID":"aad7344e-a606-4fd2-a0f3-3b87f4c9498f","Type":"ContainerStarted","Data":"914304771c99361e295955a77713afedc158bd52c56496f402ef0de96bc15cf1"} Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.856883 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fdzqc" event={"ID":"aad7344e-a606-4fd2-a0f3-3b87f4c9498f","Type":"ContainerStarted","Data":"17fa18adc4349b726223a9660458b0f19a2b5a9d403fdc2fc519e55520593c9d"} Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.863948 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hks4r" event={"ID":"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0","Type":"ContainerStarted","Data":"a1d6db17cd9aece796ec8bfbc227ee61b8f1157daff1c66794a9f4e56247fe34"} Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.867916 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-597764f7c9-2zlgs" event={"ID":"c398271d-3875-4a64-bcf3-6427e8a8762f","Type":"ContainerStarted","Data":"b6b5dddf8613b5069152e755eae503da6d84967f52045a1f140e1d7e2677aa9b"} Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.871049 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p54dp" event={"ID":"0b096b2c-0d8d-44bc-9275-e712b6e622b8","Type":"ContainerStarted","Data":"10fda8566c16914cc6daa7620204ac1ae3b0a09b92dd436053da2afadbf4e3de"} Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.877448 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tblv4" event={"ID":"ef2b1def-1464-4d5e-8d4e-44a4e505fa40","Type":"ContainerStarted","Data":"e80fc9727fd4f143a032e21ac09accab896a79494a8ede322ee1fa9ffd51da8c"} Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.878935 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-flz9d" event={"ID":"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9","Type":"ContainerStarted","Data":"92e51d8741296d37282ec649137fa18b40ffd6eee17458c0bca3394cdad5a180"} Oct 08 14:20:41 crc kubenswrapper[4789]: I1008 14:20:41.892278 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fdzqc" podStartSLOduration=2.892258185 podStartE2EDuration="2.892258185s" podCreationTimestamp="2025-10-08 14:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:20:41.886262362 +0000 UTC m=+1181.793009864" watchObservedRunningTime="2025-10-08 14:20:41.892258185 +0000 UTC m=+1181.799005667" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.289446 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:20:42 crc kubenswrapper[4789]: W1008 14:20:42.293486 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc16c0be1_2f8d_4d4e_b045_1a31a3fca002.slice/crio-e3bccd422af36397e5cfd0630258bf12891a94befcd3b24b2b3cf8c5c2b699a3 WatchSource:0}: Error finding container e3bccd422af36397e5cfd0630258bf12891a94befcd3b24b2b3cf8c5c2b699a3: Status 404 returned error can't find the container with id e3bccd422af36397e5cfd0630258bf12891a94befcd3b24b2b3cf8c5c2b699a3 Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.307015 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d7cb85455-q84nd"] Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.312041 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d44448797-vjpbw"] Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.470738 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-597764f7c9-2zlgs"] Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.544062 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d567f77d5-h9q42"] Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.545623 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.556571 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d567f77d5-h9q42"] Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.630975 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7jzn\" (UniqueName: \"kubernetes.io/projected/b3876e8c-aee2-419e-a8a8-6ed967e029b5-kube-api-access-p7jzn\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.631053 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3876e8c-aee2-419e-a8a8-6ed967e029b5-logs\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.631158 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-scripts\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.631193 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-config-data\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.631236 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3876e8c-aee2-419e-a8a8-6ed967e029b5-horizon-secret-key\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.642038 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.740699 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-scripts\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.741487 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-scripts\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.741628 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-config-data\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.742622 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-config-data\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.742709 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3876e8c-aee2-419e-a8a8-6ed967e029b5-horizon-secret-key\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.742785 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7jzn\" (UniqueName: \"kubernetes.io/projected/b3876e8c-aee2-419e-a8a8-6ed967e029b5-kube-api-access-p7jzn\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.742868 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3876e8c-aee2-419e-a8a8-6ed967e029b5-logs\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.743355 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3876e8c-aee2-419e-a8a8-6ed967e029b5-logs\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.772461 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7jzn\" (UniqueName: \"kubernetes.io/projected/b3876e8c-aee2-419e-a8a8-6ed967e029b5-kube-api-access-p7jzn\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.792419 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3876e8c-aee2-419e-a8a8-6ed967e029b5-horizon-secret-key\") pod \"horizon-d567f77d5-h9q42\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.898672 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d7cb85455-q84nd" event={"ID":"0410f74c-c9df-4218-9d63-790a093c1213","Type":"ContainerStarted","Data":"b9dceafc666e2d6d457474a12f55477087fcb32fa7cfc5cecfdbe5af4b5625ae"} Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.900856 4789 generic.go:334] "Generic (PLEG): container finished" podID="c16c0be1-2f8d-4d4e-b045-1a31a3fca002" containerID="a22f20e7f6e8dffda6ffeed1e9d6389536e12cccd3e2e29c5f7451e23e77f5f4" exitCode=0 Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.900915 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" event={"ID":"c16c0be1-2f8d-4d4e-b045-1a31a3fca002","Type":"ContainerDied","Data":"a22f20e7f6e8dffda6ffeed1e9d6389536e12cccd3e2e29c5f7451e23e77f5f4"} Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.900942 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" event={"ID":"c16c0be1-2f8d-4d4e-b045-1a31a3fca002","Type":"ContainerStarted","Data":"e3bccd422af36397e5cfd0630258bf12891a94befcd3b24b2b3cf8c5c2b699a3"} Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.906875 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p54dp" event={"ID":"0b096b2c-0d8d-44bc-9275-e712b6e622b8","Type":"ContainerStarted","Data":"c1bf52812a5d2149fb993cd3676c09b67d4e4fdde81d2f4ee339b9debb591dc0"} Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.916701 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b9fe83-3af2-413e-b7fd-cabe53206482","Type":"ContainerStarted","Data":"7f53b0e2716ce580cd9090b6275b9fc4a6b1faf7748fe0596b7225d73e0a5a9e"} Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.943812 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.950936 4789 generic.go:334] "Generic (PLEG): container finished" podID="9febac0c-7e85-4ee7-a5e8-6c65de13bda7" containerID="cdff2e70266568360f01c3538531ee0880f1f43fdc572cee6aba3a93c555ca9d" exitCode=0 Oct 08 14:20:42 crc kubenswrapper[4789]: I1008 14:20:42.955347 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" event={"ID":"9febac0c-7e85-4ee7-a5e8-6c65de13bda7","Type":"ContainerDied","Data":"cdff2e70266568360f01c3538531ee0880f1f43fdc572cee6aba3a93c555ca9d"} Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.014040 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.030109 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-p54dp" podStartSLOduration=4.030094073 podStartE2EDuration="4.030094073s" podCreationTimestamp="2025-10-08 14:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:20:42.957724701 +0000 UTC m=+1182.864472193" watchObservedRunningTime="2025-10-08 14:20:43.030094073 +0000 UTC m=+1182.936841555" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.056810 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-config\") pod \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.056906 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9wdd\" (UniqueName: \"kubernetes.io/projected/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-kube-api-access-t9wdd\") pod \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.057047 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-nb\") pod \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.057067 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-sb\") pod \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.057141 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-swift-storage-0\") pod \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.057203 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-svc\") pod \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\" (UID: \"9febac0c-7e85-4ee7-a5e8-6c65de13bda7\") " Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.063338 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-kube-api-access-t9wdd" (OuterVolumeSpecName: "kube-api-access-t9wdd") pod "9febac0c-7e85-4ee7-a5e8-6c65de13bda7" (UID: "9febac0c-7e85-4ee7-a5e8-6c65de13bda7"). InnerVolumeSpecName "kube-api-access-t9wdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.109035 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-config" (OuterVolumeSpecName: "config") pod "9febac0c-7e85-4ee7-a5e8-6c65de13bda7" (UID: "9febac0c-7e85-4ee7-a5e8-6c65de13bda7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.109666 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9febac0c-7e85-4ee7-a5e8-6c65de13bda7" (UID: "9febac0c-7e85-4ee7-a5e8-6c65de13bda7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.118085 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9febac0c-7e85-4ee7-a5e8-6c65de13bda7" (UID: "9febac0c-7e85-4ee7-a5e8-6c65de13bda7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.120467 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9febac0c-7e85-4ee7-a5e8-6c65de13bda7" (UID: "9febac0c-7e85-4ee7-a5e8-6c65de13bda7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.159242 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9wdd\" (UniqueName: \"kubernetes.io/projected/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-kube-api-access-t9wdd\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.159276 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.159306 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.159315 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.159324 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.306816 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9febac0c-7e85-4ee7-a5e8-6c65de13bda7" (UID: "9febac0c-7e85-4ee7-a5e8-6c65de13bda7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.367534 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9febac0c-7e85-4ee7-a5e8-6c65de13bda7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.524939 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d567f77d5-h9q42"] Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.969177 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" event={"ID":"c16c0be1-2f8d-4d4e-b045-1a31a3fca002","Type":"ContainerStarted","Data":"e33faf1008389661c6012ab1c47b59c3d7be4fc154f831671553bee85ceb1e59"} Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.969858 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.974429 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d567f77d5-h9q42" event={"ID":"b3876e8c-aee2-419e-a8a8-6ed967e029b5","Type":"ContainerStarted","Data":"5f3731e766323f6766bc125cc6bd616771866015e82827a797110bba7f63501b"} Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.977258 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.977255 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b7bd99c95-g2wcj" event={"ID":"9febac0c-7e85-4ee7-a5e8-6c65de13bda7","Type":"ContainerDied","Data":"452efa32bb5048a47bbbf20e298e82ca08bea4fd02817131674f92c68597d4cd"} Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.977320 4789 scope.go:117] "RemoveContainer" containerID="cdff2e70266568360f01c3538531ee0880f1f43fdc572cee6aba3a93c555ca9d" Oct 08 14:20:43 crc kubenswrapper[4789]: I1008 14:20:43.998573 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" podStartSLOduration=3.99855276 podStartE2EDuration="3.99855276s" podCreationTimestamp="2025-10-08 14:20:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:20:43.988399905 +0000 UTC m=+1183.895147417" watchObservedRunningTime="2025-10-08 14:20:43.99855276 +0000 UTC m=+1183.905300252" Oct 08 14:20:44 crc kubenswrapper[4789]: I1008 14:20:44.088738 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b7bd99c95-g2wcj"] Oct 08 14:20:44 crc kubenswrapper[4789]: I1008 14:20:44.111365 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b7bd99c95-g2wcj"] Oct 08 14:20:44 crc kubenswrapper[4789]: I1008 14:20:44.742481 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9febac0c-7e85-4ee7-a5e8-6c65de13bda7" path="/var/lib/kubelet/pods/9febac0c-7e85-4ee7-a5e8-6c65de13bda7/volumes" Oct 08 14:20:47 crc kubenswrapper[4789]: I1008 14:20:47.026530 4789 generic.go:334] "Generic (PLEG): container finished" podID="c629d8f6-586d-4f5c-a7ea-2e6912a02b03" containerID="54d9d255673843e618360f4e57b1dfe2bfe4e54a8bf1e14f224ed47b041609cc" exitCode=0 Oct 08 14:20:47 crc kubenswrapper[4789]: I1008 14:20:47.026613 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-6w7kw" event={"ID":"c629d8f6-586d-4f5c-a7ea-2e6912a02b03","Type":"ContainerDied","Data":"54d9d255673843e618360f4e57b1dfe2bfe4e54a8bf1e14f224ed47b041609cc"} Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.554314 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.683936 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-db-sync-config-data\") pod \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.684115 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-config-data\") pod \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.684222 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-combined-ca-bundle\") pod \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.684287 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vb5x\" (UniqueName: \"kubernetes.io/projected/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-kube-api-access-5vb5x\") pod \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\" (UID: \"c629d8f6-586d-4f5c-a7ea-2e6912a02b03\") " Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.692603 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-kube-api-access-5vb5x" (OuterVolumeSpecName: "kube-api-access-5vb5x") pod "c629d8f6-586d-4f5c-a7ea-2e6912a02b03" (UID: "c629d8f6-586d-4f5c-a7ea-2e6912a02b03"). InnerVolumeSpecName "kube-api-access-5vb5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.694490 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c629d8f6-586d-4f5c-a7ea-2e6912a02b03" (UID: "c629d8f6-586d-4f5c-a7ea-2e6912a02b03"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.724439 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c629d8f6-586d-4f5c-a7ea-2e6912a02b03" (UID: "c629d8f6-586d-4f5c-a7ea-2e6912a02b03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.738608 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-config-data" (OuterVolumeSpecName: "config-data") pod "c629d8f6-586d-4f5c-a7ea-2e6912a02b03" (UID: "c629d8f6-586d-4f5c-a7ea-2e6912a02b03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.787094 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.787124 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.787133 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vb5x\" (UniqueName: \"kubernetes.io/projected/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-kube-api-access-5vb5x\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:48 crc kubenswrapper[4789]: I1008 14:20:48.787144 4789 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c629d8f6-586d-4f5c-a7ea-2e6912a02b03-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.050420 4789 generic.go:334] "Generic (PLEG): container finished" podID="aad7344e-a606-4fd2-a0f3-3b87f4c9498f" containerID="914304771c99361e295955a77713afedc158bd52c56496f402ef0de96bc15cf1" exitCode=0 Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.050511 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fdzqc" event={"ID":"aad7344e-a606-4fd2-a0f3-3b87f4c9498f","Type":"ContainerDied","Data":"914304771c99361e295955a77713afedc158bd52c56496f402ef0de96bc15cf1"} Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.054669 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-6w7kw" event={"ID":"c629d8f6-586d-4f5c-a7ea-2e6912a02b03","Type":"ContainerDied","Data":"841174b907edcbd61238e44f43e55b9fe6772f86e66ed9572539dde5b384f318"} Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.054730 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="841174b907edcbd61238e44f43e55b9fe6772f86e66ed9572539dde5b384f318" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.054806 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-6w7kw" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.292820 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 08 14:20:49 crc kubenswrapper[4789]: E1008 14:20:49.293191 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9febac0c-7e85-4ee7-a5e8-6c65de13bda7" containerName="init" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.293207 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9febac0c-7e85-4ee7-a5e8-6c65de13bda7" containerName="init" Oct 08 14:20:49 crc kubenswrapper[4789]: E1008 14:20:49.293250 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c629d8f6-586d-4f5c-a7ea-2e6912a02b03" containerName="watcher-db-sync" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.293257 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c629d8f6-586d-4f5c-a7ea-2e6912a02b03" containerName="watcher-db-sync" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.293481 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9febac0c-7e85-4ee7-a5e8-6c65de13bda7" containerName="init" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.293500 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c629d8f6-586d-4f5c-a7ea-2e6912a02b03" containerName="watcher-db-sync" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.294419 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.297492 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-j7bbz" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.297641 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.309826 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.311056 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.315400 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.321882 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.350442 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.400113 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-config-data\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.400167 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7ljb\" (UniqueName: \"kubernetes.io/projected/cfe1c2a6-d542-40c0-b743-1c75caa755af-kube-api-access-m7ljb\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.400272 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.400313 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.400362 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe1c2a6-d542-40c0-b743-1c75caa755af-logs\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.457808 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.458870 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.462205 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.488640 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.527860 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.527918 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe1c2a6-d542-40c0-b743-1c75caa755af-logs\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.527962 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdtvv\" (UniqueName: \"kubernetes.io/projected/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-kube-api-access-hdtvv\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.528001 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.528050 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-config-data\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.528070 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7ljb\" (UniqueName: \"kubernetes.io/projected/cfe1c2a6-d542-40c0-b743-1c75caa755af-kube-api-access-m7ljb\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.528161 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.528190 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-config-data\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.528210 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-logs\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.547880 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.548412 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe1c2a6-d542-40c0-b743-1c75caa755af-logs\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.558829 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-config-data\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.559633 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.605458 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d7cb85455-q84nd"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.627050 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79cc7df9c4-dd4dv"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.629600 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.636180 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.641599 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7ljb\" (UniqueName: \"kubernetes.io/projected/cfe1c2a6-d542-40c0-b743-1c75caa755af-kube-api-access-m7ljb\") pod \"watcher-api-0\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.642714 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79cc7df9c4-dd4dv"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648386 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-scripts\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648452 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-config-data\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648485 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-combined-ca-bundle\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648512 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-tls-certs\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648545 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648569 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33bbfc8e-64bd-480c-b755-dd802cba53fe-logs\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648610 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-logs\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648635 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzcpl\" (UniqueName: \"kubernetes.io/projected/33bbfc8e-64bd-480c-b755-dd802cba53fe-kube-api-access-vzcpl\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648655 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-secret-key\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648710 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-config-data\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.648729 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-logs\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.649261 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f2mq\" (UniqueName: \"kubernetes.io/projected/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-kube-api-access-6f2mq\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.649318 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.649344 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.649363 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdtvv\" (UniqueName: \"kubernetes.io/projected/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-kube-api-access-hdtvv\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.649388 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-config-data\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.650114 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-logs\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.666862 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.667376 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-config-data\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.705490 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdtvv\" (UniqueName: \"kubernetes.io/projected/d84335e3-9c9e-473f-be0a-f1f4535ae6b5-kube-api-access-hdtvv\") pod \"watcher-applier-0\" (UID: \"d84335e3-9c9e-473f-be0a-f1f4535ae6b5\") " pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756063 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f2mq\" (UniqueName: \"kubernetes.io/projected/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-kube-api-access-6f2mq\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756126 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756155 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-config-data\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756188 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-scripts\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756221 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-config-data\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756243 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-combined-ca-bundle\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756266 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-tls-certs\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756292 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756315 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33bbfc8e-64bd-480c-b755-dd802cba53fe-logs\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756349 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-logs\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756366 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzcpl\" (UniqueName: \"kubernetes.io/projected/33bbfc8e-64bd-480c-b755-dd802cba53fe-kube-api-access-vzcpl\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.756382 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-secret-key\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.758539 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33bbfc8e-64bd-480c-b755-dd802cba53fe-logs\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.763841 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-config-data\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.763841 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.764479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.768745 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-logs\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.769600 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-combined-ca-bundle\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.769683 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-scripts\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.770083 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-tls-certs\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.770599 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-config-data\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.774656 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d567f77d5-h9q42"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.776367 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-secret-key\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.787786 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzcpl\" (UniqueName: \"kubernetes.io/projected/33bbfc8e-64bd-480c-b755-dd802cba53fe-kube-api-access-vzcpl\") pod \"watcher-decision-engine-0\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.795441 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.803317 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f2mq\" (UniqueName: \"kubernetes.io/projected/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-kube-api-access-6f2mq\") pod \"horizon-79cc7df9c4-dd4dv\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.825202 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6c7bdc9dd-bg9vf"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.827611 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.840783 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c7bdc9dd-bg9vf"] Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.860058 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9289b28-1fb4-47f8-aec3-3d9d616837d3-logs\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.860118 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9289b28-1fb4-47f8-aec3-3d9d616837d3-scripts\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.860152 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9289b28-1fb4-47f8-aec3-3d9d616837d3-config-data\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.860219 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9289b28-1fb4-47f8-aec3-3d9d616837d3-horizon-secret-key\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.860653 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9289b28-1fb4-47f8-aec3-3d9d616837d3-combined-ca-bundle\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.860885 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt52d\" (UniqueName: \"kubernetes.io/projected/f9289b28-1fb4-47f8-aec3-3d9d616837d3-kube-api-access-nt52d\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.860931 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9289b28-1fb4-47f8-aec3-3d9d616837d3-horizon-tls-certs\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.910457 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.929679 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.963339 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9289b28-1fb4-47f8-aec3-3d9d616837d3-combined-ca-bundle\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.963448 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt52d\" (UniqueName: \"kubernetes.io/projected/f9289b28-1fb4-47f8-aec3-3d9d616837d3-kube-api-access-nt52d\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.963477 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9289b28-1fb4-47f8-aec3-3d9d616837d3-horizon-tls-certs\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.963583 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9289b28-1fb4-47f8-aec3-3d9d616837d3-logs\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.963615 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9289b28-1fb4-47f8-aec3-3d9d616837d3-scripts\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.963641 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9289b28-1fb4-47f8-aec3-3d9d616837d3-config-data\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.963688 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9289b28-1fb4-47f8-aec3-3d9d616837d3-horizon-secret-key\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.964245 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9289b28-1fb4-47f8-aec3-3d9d616837d3-logs\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.964911 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9289b28-1fb4-47f8-aec3-3d9d616837d3-scripts\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.965746 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9289b28-1fb4-47f8-aec3-3d9d616837d3-config-data\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.969162 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9289b28-1fb4-47f8-aec3-3d9d616837d3-horizon-secret-key\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.971147 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9289b28-1fb4-47f8-aec3-3d9d616837d3-horizon-tls-certs\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.977092 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9289b28-1fb4-47f8-aec3-3d9d616837d3-combined-ca-bundle\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:49 crc kubenswrapper[4789]: I1008 14:20:49.979750 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt52d\" (UniqueName: \"kubernetes.io/projected/f9289b28-1fb4-47f8-aec3-3d9d616837d3-kube-api-access-nt52d\") pod \"horizon-6c7bdc9dd-bg9vf\" (UID: \"f9289b28-1fb4-47f8-aec3-3d9d616837d3\") " pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:50 crc kubenswrapper[4789]: I1008 14:20:50.092710 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:20:50 crc kubenswrapper[4789]: I1008 14:20:50.147145 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:20:51 crc kubenswrapper[4789]: I1008 14:20:51.027753 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:20:51 crc kubenswrapper[4789]: I1008 14:20:51.093895 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fcb49bfb9-8ch52"] Oct 08 14:20:51 crc kubenswrapper[4789]: I1008 14:20:51.094196 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" containerID="cri-o://4583edbf8887854b7713f005874fe01ccf08147dacbba0f7e06e974f3afae0b2" gracePeriod=10 Oct 08 14:20:52 crc kubenswrapper[4789]: I1008 14:20:52.102565 4789 generic.go:334] "Generic (PLEG): container finished" podID="bc8dc138-180a-44c7-817e-d3baf5231352" containerID="4583edbf8887854b7713f005874fe01ccf08147dacbba0f7e06e974f3afae0b2" exitCode=0 Oct 08 14:20:52 crc kubenswrapper[4789]: I1008 14:20:52.102661 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" event={"ID":"bc8dc138-180a-44c7-817e-d3baf5231352","Type":"ContainerDied","Data":"4583edbf8887854b7713f005874fe01ccf08147dacbba0f7e06e974f3afae0b2"} Oct 08 14:20:53 crc kubenswrapper[4789]: I1008 14:20:53.799603 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: connect: connection refused" Oct 08 14:20:58 crc kubenswrapper[4789]: I1008 14:20:58.798835 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: connect: connection refused" Oct 08 14:21:01 crc kubenswrapper[4789]: I1008 14:21:01.473611 4789 scope.go:117] "RemoveContainer" containerID="a48c1fd17b4d92f5dbdc76c90c5f8b0e6e0bdeefabe0c3bc914d597d4911aa33" Oct 08 14:21:03 crc kubenswrapper[4789]: I1008 14:21:03.799186 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: connect: connection refused" Oct 08 14:21:03 crc kubenswrapper[4789]: I1008 14:21:03.799518 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:21:05 crc kubenswrapper[4789]: E1008 14:21:05.865940 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 14:21:05 crc kubenswrapper[4789]: E1008 14:21:05.866010 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 14:21:05 crc kubenswrapper[4789]: E1008 14:21:05.866133 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8bhdbh5fdh575h5c9h56bhd9h664h56hfch5bbh58bhf9h59fh659h67dh554h96h54hdch55fh4h559hc8h646h7hf4h667h5d8h5f6h57bh56cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p7jzn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-d567f77d5-h9q42_openstack(b3876e8c-aee2-419e-a8a8-6ed967e029b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:21:05 crc kubenswrapper[4789]: E1008 14:21:05.869314 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-d567f77d5-h9q42" podUID="b3876e8c-aee2-419e-a8a8-6ed967e029b5" Oct 08 14:21:08 crc kubenswrapper[4789]: I1008 14:21:08.799530 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: connect: connection refused" Oct 08 14:21:13 crc kubenswrapper[4789]: I1008 14:21:13.800056 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: connect: connection refused" Oct 08 14:21:14 crc kubenswrapper[4789]: E1008 14:21:14.317126 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 14:21:14 crc kubenswrapper[4789]: E1008 14:21:14.317174 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 14:21:14 crc kubenswrapper[4789]: E1008 14:21:14.317287 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n59fhfdh5b5h677hf7hbbhc8h594h9fh79h576h5hbdh9fhcbh96h697hc8h566h668h557h94hfbhb5hd5h6bh94h68bh6h7dh64bh587q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bswhx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5d7cb85455-q84nd_openstack(0410f74c-c9df-4218-9d63-790a093c1213): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:21:14 crc kubenswrapper[4789]: E1008 14:21:14.322083 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-5d7cb85455-q84nd" podUID="0410f74c-c9df-4218-9d63-790a093c1213" Oct 08 14:21:15 crc kubenswrapper[4789]: E1008 14:21:15.962348 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 14:21:15 crc kubenswrapper[4789]: E1008 14:21:15.962716 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest" Oct 08 14:21:15 crc kubenswrapper[4789]: E1008 14:21:15.962835 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bch578h67fh94h5f4h66fh5f4h56h574h5b6h5f8hcch84h5cch55hd6hf4h5d8h55h564h5cbh57fh666h85h5hf9h9fh545h566h5b4h5b8hccq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f6m4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-597764f7c9-2zlgs_openstack(c398271d-3875-4a64-bcf3-6427e8a8762f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:21:15 crc kubenswrapper[4789]: E1008 14:21:15.964655 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-597764f7c9-2zlgs" podUID="c398271d-3875-4a64-bcf3-6427e8a8762f" Oct 08 14:21:16 crc kubenswrapper[4789]: E1008 14:21:16.746364 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Oct 08 14:21:16 crc kubenswrapper[4789]: E1008 14:21:16.747088 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Oct 08 14:21:16 crc kubenswrapper[4789]: E1008 14:21:16.747240 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:38.102.83.222:5001/podified-master-centos10/openstack-glance-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-txzxz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-5c9mc_openstack(5d62cf1d-f82b-4eed-8500-2614367a4e3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:21:16 crc kubenswrapper[4789]: E1008 14:21:16.748483 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-5c9mc" podUID="5d62cf1d-f82b-4eed-8500-2614367a4e3d" Oct 08 14:21:16 crc kubenswrapper[4789]: E1008 14:21:16.769790 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Oct 08 14:21:16 crc kubenswrapper[4789]: E1008 14:21:16.769860 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Oct 08 14:21:16 crc kubenswrapper[4789]: E1008 14:21:16.770015 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.102.83.222:5001/podified-master-centos10/openstack-barbican-api:watcher_latest,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4hzzj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-hks4r_openstack(dccf34c5-6e31-4d04-9478-d4e4edcc6ab0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:21:16 crc kubenswrapper[4789]: E1008 14:21:16.771522 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-hks4r" podUID="dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.860727 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.868184 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.974590 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq5hh\" (UniqueName: \"kubernetes.io/projected/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-kube-api-access-lq5hh\") pod \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.974664 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-credential-keys\") pod \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.974697 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-config-data\") pod \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.974761 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-combined-ca-bundle\") pod \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.974793 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-fernet-keys\") pod \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.974872 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3876e8c-aee2-419e-a8a8-6ed967e029b5-horizon-secret-key\") pod \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.974899 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3876e8c-aee2-419e-a8a8-6ed967e029b5-logs\") pod \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.974927 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-config-data\") pod \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.974946 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-scripts\") pod \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.975022 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-scripts\") pod \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\" (UID: \"aad7344e-a606-4fd2-a0f3-3b87f4c9498f\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.975129 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7jzn\" (UniqueName: \"kubernetes.io/projected/b3876e8c-aee2-419e-a8a8-6ed967e029b5-kube-api-access-p7jzn\") pod \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\" (UID: \"b3876e8c-aee2-419e-a8a8-6ed967e029b5\") " Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.975795 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3876e8c-aee2-419e-a8a8-6ed967e029b5-logs" (OuterVolumeSpecName: "logs") pod "b3876e8c-aee2-419e-a8a8-6ed967e029b5" (UID: "b3876e8c-aee2-419e-a8a8-6ed967e029b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.976235 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-scripts" (OuterVolumeSpecName: "scripts") pod "b3876e8c-aee2-419e-a8a8-6ed967e029b5" (UID: "b3876e8c-aee2-419e-a8a8-6ed967e029b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.976902 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-config-data" (OuterVolumeSpecName: "config-data") pod "b3876e8c-aee2-419e-a8a8-6ed967e029b5" (UID: "b3876e8c-aee2-419e-a8a8-6ed967e029b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.981438 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "aad7344e-a606-4fd2-a0f3-3b87f4c9498f" (UID: "aad7344e-a606-4fd2-a0f3-3b87f4c9498f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.982002 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3876e8c-aee2-419e-a8a8-6ed967e029b5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b3876e8c-aee2-419e-a8a8-6ed967e029b5" (UID: "b3876e8c-aee2-419e-a8a8-6ed967e029b5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.984154 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-scripts" (OuterVolumeSpecName: "scripts") pod "aad7344e-a606-4fd2-a0f3-3b87f4c9498f" (UID: "aad7344e-a606-4fd2-a0f3-3b87f4c9498f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.985164 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3876e8c-aee2-419e-a8a8-6ed967e029b5-kube-api-access-p7jzn" (OuterVolumeSpecName: "kube-api-access-p7jzn") pod "b3876e8c-aee2-419e-a8a8-6ed967e029b5" (UID: "b3876e8c-aee2-419e-a8a8-6ed967e029b5"). InnerVolumeSpecName "kube-api-access-p7jzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.985239 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "aad7344e-a606-4fd2-a0f3-3b87f4c9498f" (UID: "aad7344e-a606-4fd2-a0f3-3b87f4c9498f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:16 crc kubenswrapper[4789]: I1008 14:21:16.985481 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-kube-api-access-lq5hh" (OuterVolumeSpecName: "kube-api-access-lq5hh") pod "aad7344e-a606-4fd2-a0f3-3b87f4c9498f" (UID: "aad7344e-a606-4fd2-a0f3-3b87f4c9498f"). InnerVolumeSpecName "kube-api-access-lq5hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.007130 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aad7344e-a606-4fd2-a0f3-3b87f4c9498f" (UID: "aad7344e-a606-4fd2-a0f3-3b87f4c9498f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.023338 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-config-data" (OuterVolumeSpecName: "config-data") pod "aad7344e-a606-4fd2-a0f3-3b87f4c9498f" (UID: "aad7344e-a606-4fd2-a0f3-3b87f4c9498f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078433 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7jzn\" (UniqueName: \"kubernetes.io/projected/b3876e8c-aee2-419e-a8a8-6ed967e029b5-kube-api-access-p7jzn\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078463 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq5hh\" (UniqueName: \"kubernetes.io/projected/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-kube-api-access-lq5hh\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078497 4789 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078507 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078516 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078523 4789 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078532 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b3876e8c-aee2-419e-a8a8-6ed967e029b5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078540 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3876e8c-aee2-419e-a8a8-6ed967e029b5-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078570 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078578 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3876e8c-aee2-419e-a8a8-6ed967e029b5-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.078586 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad7344e-a606-4fd2-a0f3-3b87f4c9498f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.324065 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fdzqc" event={"ID":"aad7344e-a606-4fd2-a0f3-3b87f4c9498f","Type":"ContainerDied","Data":"17fa18adc4349b726223a9660458b0f19a2b5a9d403fdc2fc519e55520593c9d"} Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.324107 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17fa18adc4349b726223a9660458b0f19a2b5a9d403fdc2fc519e55520593c9d" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.324106 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fdzqc" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.325535 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d567f77d5-h9q42" event={"ID":"b3876e8c-aee2-419e-a8a8-6ed967e029b5","Type":"ContainerDied","Data":"5f3731e766323f6766bc125cc6bd616771866015e82827a797110bba7f63501b"} Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.325611 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d567f77d5-h9q42" Oct 08 14:21:17 crc kubenswrapper[4789]: E1008 14:21:17.326940 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-barbican-api:watcher_latest\\\"\"" pod="openstack/barbican-db-sync-hks4r" podUID="dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.406183 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d567f77d5-h9q42"] Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.417887 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d567f77d5-h9q42"] Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.942844 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fdzqc"] Oct 08 14:21:17 crc kubenswrapper[4789]: I1008 14:21:17.948835 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fdzqc"] Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.046193 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-msj9f"] Oct 08 14:21:18 crc kubenswrapper[4789]: E1008 14:21:18.046629 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad7344e-a606-4fd2-a0f3-3b87f4c9498f" containerName="keystone-bootstrap" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.046642 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad7344e-a606-4fd2-a0f3-3b87f4c9498f" containerName="keystone-bootstrap" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.046873 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad7344e-a606-4fd2-a0f3-3b87f4c9498f" containerName="keystone-bootstrap" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.047585 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.051254 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.051771 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8h2vv" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.052001 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.054074 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.057942 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-msj9f"] Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.198362 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-scripts\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.198544 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkpnf\" (UniqueName: \"kubernetes.io/projected/5f21f66b-b493-409d-aee9-fe6b97b468f5-kube-api-access-pkpnf\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.198670 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-config-data\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.198744 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-combined-ca-bundle\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.198949 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-fernet-keys\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.199196 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-credential-keys\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.301091 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-credential-keys\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.301153 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-scripts\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.301184 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkpnf\" (UniqueName: \"kubernetes.io/projected/5f21f66b-b493-409d-aee9-fe6b97b468f5-kube-api-access-pkpnf\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.301218 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-config-data\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.301248 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-combined-ca-bundle\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.301318 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-fernet-keys\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.321115 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-scripts\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.321434 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-combined-ca-bundle\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.321493 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-config-data\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.321559 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-fernet-keys\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.322150 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-credential-keys\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.325286 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkpnf\" (UniqueName: \"kubernetes.io/projected/5f21f66b-b493-409d-aee9-fe6b97b468f5-kube-api-access-pkpnf\") pod \"keystone-bootstrap-msj9f\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.370926 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.656210 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.749353 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aad7344e-a606-4fd2-a0f3-3b87f4c9498f" path="/var/lib/kubelet/pods/aad7344e-a606-4fd2-a0f3-3b87f4c9498f/volumes" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.754907 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3876e8c-aee2-419e-a8a8-6ed967e029b5" path="/var/lib/kubelet/pods/b3876e8c-aee2-419e-a8a8-6ed967e029b5/volumes" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.809105 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-config-data\") pod \"0410f74c-c9df-4218-9d63-790a093c1213\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.809167 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0410f74c-c9df-4218-9d63-790a093c1213-horizon-secret-key\") pod \"0410f74c-c9df-4218-9d63-790a093c1213\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.809193 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0410f74c-c9df-4218-9d63-790a093c1213-logs\") pod \"0410f74c-c9df-4218-9d63-790a093c1213\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.809243 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bswhx\" (UniqueName: \"kubernetes.io/projected/0410f74c-c9df-4218-9d63-790a093c1213-kube-api-access-bswhx\") pod \"0410f74c-c9df-4218-9d63-790a093c1213\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.809301 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-scripts\") pod \"0410f74c-c9df-4218-9d63-790a093c1213\" (UID: \"0410f74c-c9df-4218-9d63-790a093c1213\") " Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.809916 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0410f74c-c9df-4218-9d63-790a093c1213-logs" (OuterVolumeSpecName: "logs") pod "0410f74c-c9df-4218-9d63-790a093c1213" (UID: "0410f74c-c9df-4218-9d63-790a093c1213"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.810056 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-scripts" (OuterVolumeSpecName: "scripts") pod "0410f74c-c9df-4218-9d63-790a093c1213" (UID: "0410f74c-c9df-4218-9d63-790a093c1213"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.811004 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-config-data" (OuterVolumeSpecName: "config-data") pod "0410f74c-c9df-4218-9d63-790a093c1213" (UID: "0410f74c-c9df-4218-9d63-790a093c1213"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.814013 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0410f74c-c9df-4218-9d63-790a093c1213-kube-api-access-bswhx" (OuterVolumeSpecName: "kube-api-access-bswhx") pod "0410f74c-c9df-4218-9d63-790a093c1213" (UID: "0410f74c-c9df-4218-9d63-790a093c1213"). InnerVolumeSpecName "kube-api-access-bswhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.815125 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0410f74c-c9df-4218-9d63-790a093c1213-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0410f74c-c9df-4218-9d63-790a093c1213" (UID: "0410f74c-c9df-4218-9d63-790a093c1213"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.911528 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0410f74c-c9df-4218-9d63-790a093c1213-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.911572 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bswhx\" (UniqueName: \"kubernetes.io/projected/0410f74c-c9df-4218-9d63-790a093c1213-kube-api-access-bswhx\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.911590 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.911602 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0410f74c-c9df-4218-9d63-790a093c1213-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:18 crc kubenswrapper[4789]: I1008 14:21:18.911615 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0410f74c-c9df-4218-9d63-790a093c1213-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:19 crc kubenswrapper[4789]: E1008 14:21:19.025959 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 08 14:21:19 crc kubenswrapper[4789]: E1008 14:21:19.026041 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Oct 08 14:21:19 crc kubenswrapper[4789]: E1008 14:21:19.026186 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.222:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bs6qq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-tblv4_openstack(ef2b1def-1464-4d5e-8d4e-44a4e505fa40): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:21:19 crc kubenswrapper[4789]: E1008 14:21:19.027415 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-tblv4" podUID="ef2b1def-1464-4d5e-8d4e-44a4e505fa40" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.346297 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d7cb85455-q84nd" event={"ID":"0410f74c-c9df-4218-9d63-790a093c1213","Type":"ContainerDied","Data":"b9dceafc666e2d6d457474a12f55477087fcb32fa7cfc5cecfdbe5af4b5625ae"} Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.346357 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d7cb85455-q84nd" Oct 08 14:21:19 crc kubenswrapper[4789]: E1008 14:21:19.348614 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-tblv4" podUID="ef2b1def-1464-4d5e-8d4e-44a4e505fa40" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.406414 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d7cb85455-q84nd"] Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.416813 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5d7cb85455-q84nd"] Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.513469 4789 scope.go:117] "RemoveContainer" containerID="8136ab5aff0e5a8ccd07a6bdd7e7fe31f4dd77959188f67ee3bf69b771d31e64" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.775145 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.814625 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.936645 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6m4p\" (UniqueName: \"kubernetes.io/projected/c398271d-3875-4a64-bcf3-6427e8a8762f-kube-api-access-f6m4p\") pod \"c398271d-3875-4a64-bcf3-6427e8a8762f\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.936925 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c398271d-3875-4a64-bcf3-6427e8a8762f-horizon-secret-key\") pod \"c398271d-3875-4a64-bcf3-6427e8a8762f\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.936955 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c398271d-3875-4a64-bcf3-6427e8a8762f-logs\") pod \"c398271d-3875-4a64-bcf3-6427e8a8762f\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.937085 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-config\") pod \"bc8dc138-180a-44c7-817e-d3baf5231352\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.937112 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nd95\" (UniqueName: \"kubernetes.io/projected/bc8dc138-180a-44c7-817e-d3baf5231352-kube-api-access-9nd95\") pod \"bc8dc138-180a-44c7-817e-d3baf5231352\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.937171 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-scripts\") pod \"c398271d-3875-4a64-bcf3-6427e8a8762f\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.937219 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-sb\") pod \"bc8dc138-180a-44c7-817e-d3baf5231352\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.937247 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-svc\") pod \"bc8dc138-180a-44c7-817e-d3baf5231352\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.937276 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-config-data\") pod \"c398271d-3875-4a64-bcf3-6427e8a8762f\" (UID: \"c398271d-3875-4a64-bcf3-6427e8a8762f\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.937315 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-nb\") pod \"bc8dc138-180a-44c7-817e-d3baf5231352\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.937354 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-swift-storage-0\") pod \"bc8dc138-180a-44c7-817e-d3baf5231352\" (UID: \"bc8dc138-180a-44c7-817e-d3baf5231352\") " Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.937954 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c398271d-3875-4a64-bcf3-6427e8a8762f-logs" (OuterVolumeSpecName: "logs") pod "c398271d-3875-4a64-bcf3-6427e8a8762f" (UID: "c398271d-3875-4a64-bcf3-6427e8a8762f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.938498 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-config-data" (OuterVolumeSpecName: "config-data") pod "c398271d-3875-4a64-bcf3-6427e8a8762f" (UID: "c398271d-3875-4a64-bcf3-6427e8a8762f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.938700 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-scripts" (OuterVolumeSpecName: "scripts") pod "c398271d-3875-4a64-bcf3-6427e8a8762f" (UID: "c398271d-3875-4a64-bcf3-6427e8a8762f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.939269 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c398271d-3875-4a64-bcf3-6427e8a8762f-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.939297 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.939309 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c398271d-3875-4a64-bcf3-6427e8a8762f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.942866 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c398271d-3875-4a64-bcf3-6427e8a8762f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c398271d-3875-4a64-bcf3-6427e8a8762f" (UID: "c398271d-3875-4a64-bcf3-6427e8a8762f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.944034 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc8dc138-180a-44c7-817e-d3baf5231352-kube-api-access-9nd95" (OuterVolumeSpecName: "kube-api-access-9nd95") pod "bc8dc138-180a-44c7-817e-d3baf5231352" (UID: "bc8dc138-180a-44c7-817e-d3baf5231352"). InnerVolumeSpecName "kube-api-access-9nd95". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.944369 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c398271d-3875-4a64-bcf3-6427e8a8762f-kube-api-access-f6m4p" (OuterVolumeSpecName: "kube-api-access-f6m4p") pod "c398271d-3875-4a64-bcf3-6427e8a8762f" (UID: "c398271d-3875-4a64-bcf3-6427e8a8762f"). InnerVolumeSpecName "kube-api-access-f6m4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.984162 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bc8dc138-180a-44c7-817e-d3baf5231352" (UID: "bc8dc138-180a-44c7-817e-d3baf5231352"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.989463 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bc8dc138-180a-44c7-817e-d3baf5231352" (UID: "bc8dc138-180a-44c7-817e-d3baf5231352"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.989826 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bc8dc138-180a-44c7-817e-d3baf5231352" (UID: "bc8dc138-180a-44c7-817e-d3baf5231352"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.989981 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bc8dc138-180a-44c7-817e-d3baf5231352" (UID: "bc8dc138-180a-44c7-817e-d3baf5231352"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:19 crc kubenswrapper[4789]: I1008 14:21:19.998269 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-config" (OuterVolumeSpecName: "config") pod "bc8dc138-180a-44c7-817e-d3baf5231352" (UID: "bc8dc138-180a-44c7-817e-d3baf5231352"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.041440 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6m4p\" (UniqueName: \"kubernetes.io/projected/c398271d-3875-4a64-bcf3-6427e8a8762f-kube-api-access-f6m4p\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.041477 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c398271d-3875-4a64-bcf3-6427e8a8762f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.041487 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.041497 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nd95\" (UniqueName: \"kubernetes.io/projected/bc8dc138-180a-44c7-817e-d3baf5231352-kube-api-access-9nd95\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.041507 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.041515 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.041523 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.041532 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc8dc138-180a-44c7-817e-d3baf5231352-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.175524 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79cc7df9c4-dd4dv"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.204487 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.238347 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.358187 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-597764f7c9-2zlgs" event={"ID":"c398271d-3875-4a64-bcf3-6427e8a8762f","Type":"ContainerDied","Data":"b6b5dddf8613b5069152e755eae503da6d84967f52045a1f140e1d7e2677aa9b"} Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.360867 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"33bbfc8e-64bd-480c-b755-dd802cba53fe","Type":"ContainerStarted","Data":"5faf3fe89fcfa6973f4667c648b25e7d874fdbd51b339cda678b294c19ca4506"} Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.365295 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cfe1c2a6-d542-40c0-b743-1c75caa755af","Type":"ContainerStarted","Data":"cae3fffa2ced31b4ccaf756124735a8906a118b4a8fb9673a760673008a772c2"} Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.367630 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" event={"ID":"bc8dc138-180a-44c7-817e-d3baf5231352","Type":"ContainerDied","Data":"3664e4f95790cb3919859131f08344a1c933ca1d668c516796f7d01c71228077"} Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.367682 4789 scope.go:117] "RemoveContainer" containerID="4583edbf8887854b7713f005874fe01ccf08147dacbba0f7e06e974f3afae0b2" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.367805 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.377514 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-597764f7c9-2zlgs" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.377681 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cc7df9c4-dd4dv" event={"ID":"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0","Type":"ContainerStarted","Data":"4d9c991bf33fcb9e0e5e500b76f0ae493c1edd08105bb669fbb7830a12775e13"} Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.424445 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.433924 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c7bdc9dd-bg9vf"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.440524 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-msj9f"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.539113 4789 scope.go:117] "RemoveContainer" containerID="0a1993ad5ef13630ccb232a1839ea051ad883fce38f6cfce771177079935277d" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.580191 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fcb49bfb9-8ch52"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.588034 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fcb49bfb9-8ch52"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.613468 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-597764f7c9-2zlgs"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.626273 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-597764f7c9-2zlgs"] Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.758528 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0410f74c-c9df-4218-9d63-790a093c1213" path="/var/lib/kubelet/pods/0410f74c-c9df-4218-9d63-790a093c1213/volumes" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.758935 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" path="/var/lib/kubelet/pods/bc8dc138-180a-44c7-817e-d3baf5231352/volumes" Oct 08 14:21:20 crc kubenswrapper[4789]: I1008 14:21:20.759583 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c398271d-3875-4a64-bcf3-6427e8a8762f" path="/var/lib/kubelet/pods/c398271d-3875-4a64-bcf3-6427e8a8762f/volumes" Oct 08 14:21:21 crc kubenswrapper[4789]: I1008 14:21:21.384884 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c7bdc9dd-bg9vf" event={"ID":"f9289b28-1fb4-47f8-aec3-3d9d616837d3","Type":"ContainerStarted","Data":"dc636133d29e021d8cf29f05eb792e7568b7ea12b4f68e8183f157dc62dd40e2"} Oct 08 14:21:21 crc kubenswrapper[4789]: I1008 14:21:21.386063 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"d84335e3-9c9e-473f-be0a-f1f4535ae6b5","Type":"ContainerStarted","Data":"c4f569de98598602e3ff639079ae1f161f9c111fcae656f1ccc795e7c938a4d4"} Oct 08 14:21:21 crc kubenswrapper[4789]: I1008 14:21:21.389205 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-msj9f" event={"ID":"5f21f66b-b493-409d-aee9-fe6b97b468f5","Type":"ContainerStarted","Data":"3f2cbcdae106043e8e39c8e1305ebd2feb3f088f0f03e0099d88ba353386ca84"} Oct 08 14:21:22 crc kubenswrapper[4789]: E1008 14:21:22.813319 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Oct 08 14:21:22 crc kubenswrapper[4789]: E1008 14:21:22.813643 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" Oct 08 14:21:22 crc kubenswrapper[4789]: E1008 14:21:22.813799 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:38.102.83.222:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n55dh555h5fh5c6h5fbhb4hcbh57bh8ch687h666hddh59ch598h548hd7h8hbhf4h57ch58fh5ddh5fbhc6h8bh665h648hb8h6fh54ch57fh696q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2czbg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(40b9fe83-3af2-413e-b7fd-cabe53206482): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:21:23 crc kubenswrapper[4789]: I1008 14:21:23.414185 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cfe1c2a6-d542-40c0-b743-1c75caa755af","Type":"ContainerStarted","Data":"6c73fba0e5019c84dbd809c932b9ef0a684f49e6d651248b4bf03854715aee42"} Oct 08 14:21:23 crc kubenswrapper[4789]: I1008 14:21:23.417064 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-flz9d" event={"ID":"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9","Type":"ContainerStarted","Data":"933f8f193cc92a17c8ff9a268982d0bc59d0534b608fb4c11e4b1f07def68aa2"} Oct 08 14:21:23 crc kubenswrapper[4789]: I1008 14:21:23.420471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-msj9f" event={"ID":"5f21f66b-b493-409d-aee9-fe6b97b468f5","Type":"ContainerStarted","Data":"ef28cc9775c296870b3049d94d755a39df9b707fd45dc30f2dff0db73f8de5d4"} Oct 08 14:21:23 crc kubenswrapper[4789]: I1008 14:21:23.440617 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-flz9d" podStartSLOduration=5.638126064 podStartE2EDuration="43.44059905s" podCreationTimestamp="2025-10-08 14:20:40 +0000 UTC" firstStartedPulling="2025-10-08 14:20:41.6677363 +0000 UTC m=+1181.574483792" lastFinishedPulling="2025-10-08 14:21:19.470209286 +0000 UTC m=+1219.376956778" observedRunningTime="2025-10-08 14:21:23.434795103 +0000 UTC m=+1223.341542615" watchObservedRunningTime="2025-10-08 14:21:23.44059905 +0000 UTC m=+1223.347346542" Oct 08 14:21:23 crc kubenswrapper[4789]: I1008 14:21:23.464032 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-msj9f" podStartSLOduration=5.464014555 podStartE2EDuration="5.464014555s" podCreationTimestamp="2025-10-08 14:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:21:23.457576171 +0000 UTC m=+1223.364323673" watchObservedRunningTime="2025-10-08 14:21:23.464014555 +0000 UTC m=+1223.370762047" Oct 08 14:21:23 crc kubenswrapper[4789]: I1008 14:21:23.799426 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5fcb49bfb9-8ch52" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: i/o timeout" Oct 08 14:21:24 crc kubenswrapper[4789]: I1008 14:21:24.431165 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cfe1c2a6-d542-40c0-b743-1c75caa755af","Type":"ContainerStarted","Data":"21f184de5dd2c5c0e9e571f38377f7eea2ca82e938e161eee5a2202be40b2e27"} Oct 08 14:21:24 crc kubenswrapper[4789]: I1008 14:21:24.431521 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 08 14:21:24 crc kubenswrapper[4789]: I1008 14:21:24.436146 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cc7df9c4-dd4dv" event={"ID":"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0","Type":"ContainerStarted","Data":"3759762f30e8117b005cedf97abf2332be53ee3440a1616479c3c9aa5ed5b7c7"} Oct 08 14:21:24 crc kubenswrapper[4789]: I1008 14:21:24.436203 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cc7df9c4-dd4dv" event={"ID":"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0","Type":"ContainerStarted","Data":"b9f99165330ce21d34a9c1a6024be68840c5bbdb942c92c790f1bada38c7a935"} Oct 08 14:21:24 crc kubenswrapper[4789]: I1008 14:21:24.440741 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c7bdc9dd-bg9vf" event={"ID":"f9289b28-1fb4-47f8-aec3-3d9d616837d3","Type":"ContainerStarted","Data":"78d0fdefaeadcd0d39933bf3e819be7f414540529896769a88c1a3f15dec5e6c"} Oct 08 14:21:24 crc kubenswrapper[4789]: I1008 14:21:24.453072 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=35.45305592 podStartE2EDuration="35.45305592s" podCreationTimestamp="2025-10-08 14:20:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:21:24.450296555 +0000 UTC m=+1224.357044057" watchObservedRunningTime="2025-10-08 14:21:24.45305592 +0000 UTC m=+1224.359803402" Oct 08 14:21:24 crc kubenswrapper[4789]: I1008 14:21:24.482238 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79cc7df9c4-dd4dv" podStartSLOduration=32.357555626999996 podStartE2EDuration="35.482220231s" podCreationTimestamp="2025-10-08 14:20:49 +0000 UTC" firstStartedPulling="2025-10-08 14:21:20.222195016 +0000 UTC m=+1220.128942508" lastFinishedPulling="2025-10-08 14:21:23.34685962 +0000 UTC m=+1223.253607112" observedRunningTime="2025-10-08 14:21:24.477630216 +0000 UTC m=+1224.384377728" watchObservedRunningTime="2025-10-08 14:21:24.482220231 +0000 UTC m=+1224.388967723" Oct 08 14:21:24 crc kubenswrapper[4789]: I1008 14:21:24.911055 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 08 14:21:26 crc kubenswrapper[4789]: I1008 14:21:26.484798 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c7bdc9dd-bg9vf" event={"ID":"f9289b28-1fb4-47f8-aec3-3d9d616837d3","Type":"ContainerStarted","Data":"606654e4268863603609d3a1d045800a36bad5fa35f07c395d4ea1b0fd23c67d"} Oct 08 14:21:26 crc kubenswrapper[4789]: I1008 14:21:26.498702 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"d84335e3-9c9e-473f-be0a-f1f4535ae6b5","Type":"ContainerStarted","Data":"85d2465a9344207fcb94ec6a4b19ec5fae01f01cfe4b1688bb1f8afc1acf6f46"} Oct 08 14:21:26 crc kubenswrapper[4789]: I1008 14:21:26.505821 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6c7bdc9dd-bg9vf" podStartSLOduration=34.342758719 podStartE2EDuration="37.505802743s" podCreationTimestamp="2025-10-08 14:20:49 +0000 UTC" firstStartedPulling="2025-10-08 14:21:20.421704723 +0000 UTC m=+1220.328452215" lastFinishedPulling="2025-10-08 14:21:23.584748747 +0000 UTC m=+1223.491496239" observedRunningTime="2025-10-08 14:21:26.502917465 +0000 UTC m=+1226.409664967" watchObservedRunningTime="2025-10-08 14:21:26.505802743 +0000 UTC m=+1226.412550235" Oct 08 14:21:26 crc kubenswrapper[4789]: I1008 14:21:26.509542 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:21:26 crc kubenswrapper[4789]: I1008 14:21:26.510048 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b9fe83-3af2-413e-b7fd-cabe53206482","Type":"ContainerStarted","Data":"7490defa0c96cd57eefcb7a504e9e31cf7452f614d3e942bcb01b991fcbe0ad8"} Oct 08 14:21:26 crc kubenswrapper[4789]: I1008 14:21:26.537350 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=32.378206396 podStartE2EDuration="37.537329507s" podCreationTimestamp="2025-10-08 14:20:49 +0000 UTC" firstStartedPulling="2025-10-08 14:21:20.243151694 +0000 UTC m=+1220.149899186" lastFinishedPulling="2025-10-08 14:21:25.402274805 +0000 UTC m=+1225.309022297" observedRunningTime="2025-10-08 14:21:26.531499619 +0000 UTC m=+1226.438247111" watchObservedRunningTime="2025-10-08 14:21:26.537329507 +0000 UTC m=+1226.444076999" Oct 08 14:21:26 crc kubenswrapper[4789]: I1008 14:21:26.552203 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=32.480713045 podStartE2EDuration="37.55218095s" podCreationTimestamp="2025-10-08 14:20:49 +0000 UTC" firstStartedPulling="2025-10-08 14:21:20.420745587 +0000 UTC m=+1220.327493079" lastFinishedPulling="2025-10-08 14:21:25.492213492 +0000 UTC m=+1225.398960984" observedRunningTime="2025-10-08 14:21:26.545954701 +0000 UTC m=+1226.452702193" watchObservedRunningTime="2025-10-08 14:21:26.55218095 +0000 UTC m=+1226.458928442" Oct 08 14:21:27 crc kubenswrapper[4789]: I1008 14:21:27.033372 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 08 14:21:27 crc kubenswrapper[4789]: I1008 14:21:27.518509 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"33bbfc8e-64bd-480c-b755-dd802cba53fe","Type":"ContainerStarted","Data":"2243dc665e6aa408ef5e8d4e4ede44e6f00de22af3a81a8d794d074e1b3df648"} Oct 08 14:21:27 crc kubenswrapper[4789]: E1008 14:21:27.732625 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-glance-api:watcher_latest\\\"\"" pod="openstack/glance-db-sync-5c9mc" podUID="5d62cf1d-f82b-4eed-8500-2614367a4e3d" Oct 08 14:21:29 crc kubenswrapper[4789]: I1008 14:21:29.796434 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 14:21:29 crc kubenswrapper[4789]: I1008 14:21:29.823265 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 08 14:21:29 crc kubenswrapper[4789]: I1008 14:21:29.911057 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 08 14:21:29 crc kubenswrapper[4789]: I1008 14:21:29.923388 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 08 14:21:29 crc kubenswrapper[4789]: I1008 14:21:29.930529 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Oct 08 14:21:29 crc kubenswrapper[4789]: I1008 14:21:29.930573 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Oct 08 14:21:29 crc kubenswrapper[4789]: I1008 14:21:29.962733 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Oct 08 14:21:30 crc kubenswrapper[4789]: I1008 14:21:30.093372 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:21:30 crc kubenswrapper[4789]: I1008 14:21:30.093742 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:21:30 crc kubenswrapper[4789]: I1008 14:21:30.148189 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:21:30 crc kubenswrapper[4789]: I1008 14:21:30.148305 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:21:30 crc kubenswrapper[4789]: I1008 14:21:30.558907 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 14:21:30 crc kubenswrapper[4789]: I1008 14:21:30.566000 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 08 14:21:30 crc kubenswrapper[4789]: I1008 14:21:30.589485 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Oct 08 14:21:30 crc kubenswrapper[4789]: I1008 14:21:30.601528 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 08 14:21:33 crc kubenswrapper[4789]: I1008 14:21:33.133387 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 08 14:21:33 crc kubenswrapper[4789]: I1008 14:21:33.133958 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api-log" containerID="cri-o://6c73fba0e5019c84dbd809c932b9ef0a684f49e6d651248b4bf03854715aee42" gracePeriod=30 Oct 08 14:21:33 crc kubenswrapper[4789]: I1008 14:21:33.134471 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api" containerID="cri-o://21f184de5dd2c5c0e9e571f38377f7eea2ca82e938e161eee5a2202be40b2e27" gracePeriod=30 Oct 08 14:21:33 crc kubenswrapper[4789]: I1008 14:21:33.591069 4789 generic.go:334] "Generic (PLEG): container finished" podID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerID="6c73fba0e5019c84dbd809c932b9ef0a684f49e6d651248b4bf03854715aee42" exitCode=143 Oct 08 14:21:33 crc kubenswrapper[4789]: I1008 14:21:33.591166 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cfe1c2a6-d542-40c0-b743-1c75caa755af","Type":"ContainerDied","Data":"6c73fba0e5019c84dbd809c932b9ef0a684f49e6d651248b4bf03854715aee42"} Oct 08 14:21:34 crc kubenswrapper[4789]: I1008 14:21:34.601635 4789 generic.go:334] "Generic (PLEG): container finished" podID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerID="21f184de5dd2c5c0e9e571f38377f7eea2ca82e938e161eee5a2202be40b2e27" exitCode=0 Oct 08 14:21:34 crc kubenswrapper[4789]: I1008 14:21:34.601888 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cfe1c2a6-d542-40c0-b743-1c75caa755af","Type":"ContainerDied","Data":"21f184de5dd2c5c0e9e571f38377f7eea2ca82e938e161eee5a2202be40b2e27"} Oct 08 14:21:34 crc kubenswrapper[4789]: I1008 14:21:34.911289 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.159:9322/\": dial tcp 10.217.0.159:9322: connect: connection refused" Oct 08 14:21:34 crc kubenswrapper[4789]: I1008 14:21:34.911343 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9322/\": dial tcp 10.217.0.159:9322: connect: connection refused" Oct 08 14:21:37 crc kubenswrapper[4789]: I1008 14:21:37.923164 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.010617 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7ljb\" (UniqueName: \"kubernetes.io/projected/cfe1c2a6-d542-40c0-b743-1c75caa755af-kube-api-access-m7ljb\") pod \"cfe1c2a6-d542-40c0-b743-1c75caa755af\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.010727 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe1c2a6-d542-40c0-b743-1c75caa755af-logs\") pod \"cfe1c2a6-d542-40c0-b743-1c75caa755af\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.010859 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-custom-prometheus-ca\") pod \"cfe1c2a6-d542-40c0-b743-1c75caa755af\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.010905 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-config-data\") pod \"cfe1c2a6-d542-40c0-b743-1c75caa755af\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.010931 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-combined-ca-bundle\") pod \"cfe1c2a6-d542-40c0-b743-1c75caa755af\" (UID: \"cfe1c2a6-d542-40c0-b743-1c75caa755af\") " Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.012226 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfe1c2a6-d542-40c0-b743-1c75caa755af-logs" (OuterVolumeSpecName: "logs") pod "cfe1c2a6-d542-40c0-b743-1c75caa755af" (UID: "cfe1c2a6-d542-40c0-b743-1c75caa755af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.025874 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe1c2a6-d542-40c0-b743-1c75caa755af-kube-api-access-m7ljb" (OuterVolumeSpecName: "kube-api-access-m7ljb") pod "cfe1c2a6-d542-40c0-b743-1c75caa755af" (UID: "cfe1c2a6-d542-40c0-b743-1c75caa755af"). InnerVolumeSpecName "kube-api-access-m7ljb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.051117 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfe1c2a6-d542-40c0-b743-1c75caa755af" (UID: "cfe1c2a6-d542-40c0-b743-1c75caa755af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.072096 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-config-data" (OuterVolumeSpecName: "config-data") pod "cfe1c2a6-d542-40c0-b743-1c75caa755af" (UID: "cfe1c2a6-d542-40c0-b743-1c75caa755af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.076553 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "cfe1c2a6-d542-40c0-b743-1c75caa755af" (UID: "cfe1c2a6-d542-40c0-b743-1c75caa755af"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.113218 4789 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.113485 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.113555 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe1c2a6-d542-40c0-b743-1c75caa755af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.113643 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7ljb\" (UniqueName: \"kubernetes.io/projected/cfe1c2a6-d542-40c0-b743-1c75caa755af-kube-api-access-m7ljb\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.113723 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe1c2a6-d542-40c0-b743-1c75caa755af-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.635443 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"cfe1c2a6-d542-40c0-b743-1c75caa755af","Type":"ContainerDied","Data":"cae3fffa2ced31b4ccaf756124735a8906a118b4a8fb9673a760673008a772c2"} Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.635755 4789 scope.go:117] "RemoveContainer" containerID="21f184de5dd2c5c0e9e571f38377f7eea2ca82e938e161eee5a2202be40b2e27" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.635667 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.657439 4789 scope.go:117] "RemoveContainer" containerID="6c73fba0e5019c84dbd809c932b9ef0a684f49e6d651248b4bf03854715aee42" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.677483 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.715950 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.744080 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" path="/var/lib/kubelet/pods/cfe1c2a6-d542-40c0-b743-1c75caa755af/volumes" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.744757 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 08 14:21:38 crc kubenswrapper[4789]: E1008 14:21:38.746411 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="init" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.746438 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="init" Oct 08 14:21:38 crc kubenswrapper[4789]: E1008 14:21:38.746496 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.746505 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" Oct 08 14:21:38 crc kubenswrapper[4789]: E1008 14:21:38.746551 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.746562 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api" Oct 08 14:21:38 crc kubenswrapper[4789]: E1008 14:21:38.746580 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api-log" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.746588 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api-log" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.746895 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.746921 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe1c2a6-d542-40c0-b743-1c75caa755af" containerName="watcher-api-log" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.746940 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc8dc138-180a-44c7-817e-d3baf5231352" containerName="dnsmasq-dns" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.748145 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.748308 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.754917 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.755175 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.755459 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.826504 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-logs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.826716 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-public-tls-certs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.826970 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dswjs\" (UniqueName: \"kubernetes.io/projected/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-kube-api-access-dswjs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.827090 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.827120 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.827148 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.827192 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-config-data\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.928623 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dswjs\" (UniqueName: \"kubernetes.io/projected/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-kube-api-access-dswjs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.928710 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.928743 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.928785 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.928846 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-config-data\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.928902 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-logs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.928955 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-public-tls-certs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.929474 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-logs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.933852 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-public-tls-certs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.934216 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-config-data\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.934400 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.934711 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.948839 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:38 crc kubenswrapper[4789]: I1008 14:21:38.950180 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dswjs\" (UniqueName: \"kubernetes.io/projected/4bdbb9f8-14b6-4f34-b818-a9deb84247cf-kube-api-access-dswjs\") pod \"watcher-api-0\" (UID: \"4bdbb9f8-14b6-4f34-b818-a9deb84247cf\") " pod="openstack/watcher-api-0" Oct 08 14:21:39 crc kubenswrapper[4789]: I1008 14:21:39.070308 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 08 14:21:39 crc kubenswrapper[4789]: W1008 14:21:39.495175 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bdbb9f8_14b6_4f34_b818_a9deb84247cf.slice/crio-2f8ad70d301bc1f8f56df72cd94cd6ef19f8de953741a29ce1097caf60e47705 WatchSource:0}: Error finding container 2f8ad70d301bc1f8f56df72cd94cd6ef19f8de953741a29ce1097caf60e47705: Status 404 returned error can't find the container with id 2f8ad70d301bc1f8f56df72cd94cd6ef19f8de953741a29ce1097caf60e47705 Oct 08 14:21:39 crc kubenswrapper[4789]: I1008 14:21:39.505350 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 08 14:21:39 crc kubenswrapper[4789]: I1008 14:21:39.648869 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"4bdbb9f8-14b6-4f34-b818-a9deb84247cf","Type":"ContainerStarted","Data":"2f8ad70d301bc1f8f56df72cd94cd6ef19f8de953741a29ce1097caf60e47705"} Oct 08 14:21:40 crc kubenswrapper[4789]: I1008 14:21:40.152504 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6c7bdc9dd-bg9vf" podUID="f9289b28-1fb4-47f8-aec3-3d9d616837d3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.163:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.163:8443: connect: connection refused" Oct 08 14:21:42 crc kubenswrapper[4789]: I1008 14:21:42.008319 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:21:42 crc kubenswrapper[4789]: I1008 14:21:42.674806 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"4bdbb9f8-14b6-4f34-b818-a9deb84247cf","Type":"ContainerStarted","Data":"91f8f20bcb7b03629a7edb262676af50e9739a68e5c82d6903ea2de97d5ab4ad"} Oct 08 14:21:43 crc kubenswrapper[4789]: I1008 14:21:43.683667 4789 generic.go:334] "Generic (PLEG): container finished" podID="5f21f66b-b493-409d-aee9-fe6b97b468f5" containerID="ef28cc9775c296870b3049d94d755a39df9b707fd45dc30f2dff0db73f8de5d4" exitCode=0 Oct 08 14:21:43 crc kubenswrapper[4789]: I1008 14:21:43.683862 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-msj9f" event={"ID":"5f21f66b-b493-409d-aee9-fe6b97b468f5","Type":"ContainerDied","Data":"ef28cc9775c296870b3049d94d755a39df9b707fd45dc30f2dff0db73f8de5d4"} Oct 08 14:21:43 crc kubenswrapper[4789]: I1008 14:21:43.745320 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:21:44 crc kubenswrapper[4789]: I1008 14:21:44.695174 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"4bdbb9f8-14b6-4f34-b818-a9deb84247cf","Type":"ContainerStarted","Data":"eb9f468a39572bf5991aa96a404dd32df289cb46311bf217681f2de28a1adda0"} Oct 08 14:21:44 crc kubenswrapper[4789]: I1008 14:21:44.738092 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=6.7380601890000005 podStartE2EDuration="6.738060189s" podCreationTimestamp="2025-10-08 14:21:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:21:44.721141921 +0000 UTC m=+1244.627889423" watchObservedRunningTime="2025-10-08 14:21:44.738060189 +0000 UTC m=+1244.644807671" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.122861 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.258451 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-credential-keys\") pod \"5f21f66b-b493-409d-aee9-fe6b97b468f5\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.258505 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-fernet-keys\") pod \"5f21f66b-b493-409d-aee9-fe6b97b468f5\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.258606 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkpnf\" (UniqueName: \"kubernetes.io/projected/5f21f66b-b493-409d-aee9-fe6b97b468f5-kube-api-access-pkpnf\") pod \"5f21f66b-b493-409d-aee9-fe6b97b468f5\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.258694 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-config-data\") pod \"5f21f66b-b493-409d-aee9-fe6b97b468f5\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.258733 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-scripts\") pod \"5f21f66b-b493-409d-aee9-fe6b97b468f5\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.258756 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-combined-ca-bundle\") pod \"5f21f66b-b493-409d-aee9-fe6b97b468f5\" (UID: \"5f21f66b-b493-409d-aee9-fe6b97b468f5\") " Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.264671 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5f21f66b-b493-409d-aee9-fe6b97b468f5" (UID: "5f21f66b-b493-409d-aee9-fe6b97b468f5"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.265239 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f21f66b-b493-409d-aee9-fe6b97b468f5-kube-api-access-pkpnf" (OuterVolumeSpecName: "kube-api-access-pkpnf") pod "5f21f66b-b493-409d-aee9-fe6b97b468f5" (UID: "5f21f66b-b493-409d-aee9-fe6b97b468f5"). InnerVolumeSpecName "kube-api-access-pkpnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.265312 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5f21f66b-b493-409d-aee9-fe6b97b468f5" (UID: "5f21f66b-b493-409d-aee9-fe6b97b468f5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.266496 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-scripts" (OuterVolumeSpecName: "scripts") pod "5f21f66b-b493-409d-aee9-fe6b97b468f5" (UID: "5f21f66b-b493-409d-aee9-fe6b97b468f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.290314 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f21f66b-b493-409d-aee9-fe6b97b468f5" (UID: "5f21f66b-b493-409d-aee9-fe6b97b468f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.290726 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-config-data" (OuterVolumeSpecName: "config-data") pod "5f21f66b-b493-409d-aee9-fe6b97b468f5" (UID: "5f21f66b-b493-409d-aee9-fe6b97b468f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.360975 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkpnf\" (UniqueName: \"kubernetes.io/projected/5f21f66b-b493-409d-aee9-fe6b97b468f5-kube-api-access-pkpnf\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.361035 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.361048 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.361057 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.361067 4789 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.361081 4789 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f21f66b-b493-409d-aee9-fe6b97b468f5-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.706922 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-msj9f" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.707146 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-msj9f" event={"ID":"5f21f66b-b493-409d-aee9-fe6b97b468f5","Type":"ContainerDied","Data":"3f2cbcdae106043e8e39c8e1305ebd2feb3f088f0f03e0099d88ba353386ca84"} Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.707414 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f2cbcdae106043e8e39c8e1305ebd2feb3f088f0f03e0099d88ba353386ca84" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.708317 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 08 14:21:45 crc kubenswrapper[4789]: I1008 14:21:45.710409 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="4bdbb9f8-14b6-4f34-b818-a9deb84247cf" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.165:9322/\": dial tcp 10.217.0.165:9322: connect: connection refused" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.143255 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5c9c77979c-fl8sh"] Oct 08 14:21:46 crc kubenswrapper[4789]: E1008 14:21:46.143683 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f21f66b-b493-409d-aee9-fe6b97b468f5" containerName="keystone-bootstrap" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.143700 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f21f66b-b493-409d-aee9-fe6b97b468f5" containerName="keystone-bootstrap" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.143881 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f21f66b-b493-409d-aee9-fe6b97b468f5" containerName="keystone-bootstrap" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.144472 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.147599 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.148184 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.148248 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8h2vv" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.148752 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.149063 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.149245 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.156015 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c9c77979c-fl8sh"] Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.283627 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-combined-ca-bundle\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.283716 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-config-data\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.283889 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwznh\" (UniqueName: \"kubernetes.io/projected/5eeed1f7-b400-4c25-851d-43a1f036fa8b-kube-api-access-zwznh\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.283963 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-fernet-keys\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.284050 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-credential-keys\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.284193 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-public-tls-certs\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.284292 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-scripts\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.284328 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-internal-tls-certs\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.385883 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-internal-tls-certs\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.385937 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-combined-ca-bundle\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.385975 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-config-data\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.386053 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwznh\" (UniqueName: \"kubernetes.io/projected/5eeed1f7-b400-4c25-851d-43a1f036fa8b-kube-api-access-zwznh\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.386089 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-fernet-keys\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.386112 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-credential-keys\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.386206 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-public-tls-certs\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.386266 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-scripts\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.390629 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-internal-tls-certs\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.391413 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-fernet-keys\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.392687 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-config-data\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.393835 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-combined-ca-bundle\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.394397 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-scripts\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.401767 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-public-tls-certs\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.404877 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5eeed1f7-b400-4c25-851d-43a1f036fa8b-credential-keys\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.405763 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwznh\" (UniqueName: \"kubernetes.io/projected/5eeed1f7-b400-4c25-851d-43a1f036fa8b-kube-api-access-zwznh\") pod \"keystone-5c9c77979c-fl8sh\" (UID: \"5eeed1f7-b400-4c25-851d-43a1f036fa8b\") " pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:46 crc kubenswrapper[4789]: I1008 14:21:46.472040 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:49 crc kubenswrapper[4789]: I1008 14:21:49.070815 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 08 14:21:49 crc kubenswrapper[4789]: I1008 14:21:49.071164 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 08 14:21:50 crc kubenswrapper[4789]: I1008 14:21:50.075181 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/watcher-api-0" podUID="4bdbb9f8-14b6-4f34-b818-a9deb84247cf" containerName="watcher-api-log" probeResult="failure" output="Get \"https://10.217.0.165:9322/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:21:51 crc kubenswrapper[4789]: I1008 14:21:51.718136 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="4bdbb9f8-14b6-4f34-b818-a9deb84247cf" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.165:9322/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:21:51 crc kubenswrapper[4789]: I1008 14:21:51.929092 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 08 14:21:52 crc kubenswrapper[4789]: I1008 14:21:52.638305 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:21:54 crc kubenswrapper[4789]: I1008 14:21:54.716934 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6c7bdc9dd-bg9vf" Oct 08 14:21:54 crc kubenswrapper[4789]: I1008 14:21:54.810719 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79cc7df9c4-dd4dv"] Oct 08 14:21:54 crc kubenswrapper[4789]: I1008 14:21:54.810981 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79cc7df9c4-dd4dv" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon-log" containerID="cri-o://b9f99165330ce21d34a9c1a6024be68840c5bbdb942c92c790f1bada38c7a935" gracePeriod=30 Oct 08 14:21:54 crc kubenswrapper[4789]: I1008 14:21:54.811043 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79cc7df9c4-dd4dv" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon" containerID="cri-o://3759762f30e8117b005cedf97abf2332be53ee3440a1616479c3c9aa5ed5b7c7" gracePeriod=30 Oct 08 14:21:55 crc kubenswrapper[4789]: I1008 14:21:55.794165 4789 generic.go:334] "Generic (PLEG): container finished" podID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerID="3759762f30e8117b005cedf97abf2332be53ee3440a1616479c3c9aa5ed5b7c7" exitCode=0 Oct 08 14:21:55 crc kubenswrapper[4789]: I1008 14:21:55.794221 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cc7df9c4-dd4dv" event={"ID":"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0","Type":"ContainerDied","Data":"3759762f30e8117b005cedf97abf2332be53ee3440a1616479c3c9aa5ed5b7c7"} Oct 08 14:21:56 crc kubenswrapper[4789]: I1008 14:21:56.643409 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 14:21:56 crc kubenswrapper[4789]: I1008 14:21:56.678211 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c9c77979c-fl8sh"] Oct 08 14:21:56 crc kubenswrapper[4789]: W1008 14:21:56.683174 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5eeed1f7_b400_4c25_851d_43a1f036fa8b.slice/crio-a6698237fa5f5fc875a9abc21578846745792ac528053dfd70599817d0be355f WatchSource:0}: Error finding container a6698237fa5f5fc875a9abc21578846745792ac528053dfd70599817d0be355f: Status 404 returned error can't find the container with id a6698237fa5f5fc875a9abc21578846745792ac528053dfd70599817d0be355f Oct 08 14:21:56 crc kubenswrapper[4789]: I1008 14:21:56.807100 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b9fe83-3af2-413e-b7fd-cabe53206482","Type":"ContainerStarted","Data":"9d6eb57b02ece30ff5c215cbb5e23554bda0103123705a057368a3d0a1c0212c"} Oct 08 14:21:56 crc kubenswrapper[4789]: I1008 14:21:56.808270 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c9c77979c-fl8sh" event={"ID":"5eeed1f7-b400-4c25-851d-43a1f036fa8b","Type":"ContainerStarted","Data":"a6698237fa5f5fc875a9abc21578846745792ac528053dfd70599817d0be355f"} Oct 08 14:21:56 crc kubenswrapper[4789]: I1008 14:21:56.810594 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hks4r" event={"ID":"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0","Type":"ContainerStarted","Data":"c6a9a9f5178d5155c28bdee0e913b49dcfa8d0018b023832fff25943780f9225"} Oct 08 14:21:56 crc kubenswrapper[4789]: I1008 14:21:56.863613 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-hks4r" podStartSLOduration=2.369617088 podStartE2EDuration="1m16.863592515s" podCreationTimestamp="2025-10-08 14:20:40 +0000 UTC" firstStartedPulling="2025-10-08 14:20:41.734632483 +0000 UTC m=+1181.641379975" lastFinishedPulling="2025-10-08 14:21:56.22860791 +0000 UTC m=+1256.135355402" observedRunningTime="2025-10-08 14:21:56.861356044 +0000 UTC m=+1256.768103536" watchObservedRunningTime="2025-10-08 14:21:56.863592515 +0000 UTC m=+1256.770340007" Oct 08 14:21:57 crc kubenswrapper[4789]: I1008 14:21:57.821833 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c9c77979c-fl8sh" event={"ID":"5eeed1f7-b400-4c25-851d-43a1f036fa8b","Type":"ContainerStarted","Data":"7f582e3ffaef65895518fe8f9bad214695be702c708e580b20e219b8690d4edf"} Oct 08 14:21:57 crc kubenswrapper[4789]: I1008 14:21:57.822206 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:21:57 crc kubenswrapper[4789]: I1008 14:21:57.826546 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5c9mc" event={"ID":"5d62cf1d-f82b-4eed-8500-2614367a4e3d","Type":"ContainerStarted","Data":"25383f9b103897430455bdc8eb3b2c62de60a49ab00c45d36fae689fc954e7c7"} Oct 08 14:21:57 crc kubenswrapper[4789]: I1008 14:21:57.829535 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tblv4" event={"ID":"ef2b1def-1464-4d5e-8d4e-44a4e505fa40","Type":"ContainerStarted","Data":"0b0da7215214f8a92093bc4106a2506ee926eac0474f267783e915ef17c69b45"} Oct 08 14:21:57 crc kubenswrapper[4789]: I1008 14:21:57.860360 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5c9c77979c-fl8sh" podStartSLOduration=11.860335334 podStartE2EDuration="11.860335334s" podCreationTimestamp="2025-10-08 14:21:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:21:57.845196934 +0000 UTC m=+1257.751944436" watchObservedRunningTime="2025-10-08 14:21:57.860335334 +0000 UTC m=+1257.767082836" Oct 08 14:21:57 crc kubenswrapper[4789]: I1008 14:21:57.880005 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-tblv4" podStartSLOduration=3.384132449 podStartE2EDuration="1m17.879966026s" podCreationTimestamp="2025-10-08 14:20:40 +0000 UTC" firstStartedPulling="2025-10-08 14:20:41.734633733 +0000 UTC m=+1181.641381225" lastFinishedPulling="2025-10-08 14:21:56.23046729 +0000 UTC m=+1256.137214802" observedRunningTime="2025-10-08 14:21:57.865660588 +0000 UTC m=+1257.772408080" watchObservedRunningTime="2025-10-08 14:21:57.879966026 +0000 UTC m=+1257.786713518" Oct 08 14:21:57 crc kubenswrapper[4789]: I1008 14:21:57.901910 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-5c9mc" podStartSLOduration=3.037671712 podStartE2EDuration="1m37.90189007s" podCreationTimestamp="2025-10-08 14:20:20 +0000 UTC" firstStartedPulling="2025-10-08 14:20:21.358591575 +0000 UTC m=+1161.265339067" lastFinishedPulling="2025-10-08 14:21:56.222809933 +0000 UTC m=+1256.129557425" observedRunningTime="2025-10-08 14:21:57.891754426 +0000 UTC m=+1257.798501938" watchObservedRunningTime="2025-10-08 14:21:57.90189007 +0000 UTC m=+1257.808637562" Oct 08 14:21:59 crc kubenswrapper[4789]: I1008 14:21:59.081636 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 08 14:21:59 crc kubenswrapper[4789]: I1008 14:21:59.089631 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 08 14:22:00 crc kubenswrapper[4789]: I1008 14:22:00.094786 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79cc7df9c4-dd4dv" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.162:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.162:8443: connect: connection refused" Oct 08 14:22:05 crc kubenswrapper[4789]: E1008 14:22:05.473620 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" Oct 08 14:22:05 crc kubenswrapper[4789]: I1008 14:22:05.922175 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b9fe83-3af2-413e-b7fd-cabe53206482","Type":"ContainerStarted","Data":"9b0bfc38415de3bbdb2858d0bf6ed7e183f36528d2efaca79d6fb9d25ebd343c"} Oct 08 14:22:05 crc kubenswrapper[4789]: I1008 14:22:05.922539 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="ceilometer-notification-agent" containerID="cri-o://7490defa0c96cd57eefcb7a504e9e31cf7452f614d3e942bcb01b991fcbe0ad8" gracePeriod=30 Oct 08 14:22:05 crc kubenswrapper[4789]: I1008 14:22:05.922867 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 14:22:05 crc kubenswrapper[4789]: I1008 14:22:05.923152 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="proxy-httpd" containerID="cri-o://9b0bfc38415de3bbdb2858d0bf6ed7e183f36528d2efaca79d6fb9d25ebd343c" gracePeriod=30 Oct 08 14:22:05 crc kubenswrapper[4789]: I1008 14:22:05.923239 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="sg-core" containerID="cri-o://9d6eb57b02ece30ff5c215cbb5e23554bda0103123705a057368a3d0a1c0212c" gracePeriod=30 Oct 08 14:22:06 crc kubenswrapper[4789]: I1008 14:22:06.948806 4789 generic.go:334] "Generic (PLEG): container finished" podID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerID="9b0bfc38415de3bbdb2858d0bf6ed7e183f36528d2efaca79d6fb9d25ebd343c" exitCode=0 Oct 08 14:22:06 crc kubenswrapper[4789]: I1008 14:22:06.948847 4789 generic.go:334] "Generic (PLEG): container finished" podID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerID="9d6eb57b02ece30ff5c215cbb5e23554bda0103123705a057368a3d0a1c0212c" exitCode=2 Oct 08 14:22:06 crc kubenswrapper[4789]: I1008 14:22:06.948868 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b9fe83-3af2-413e-b7fd-cabe53206482","Type":"ContainerDied","Data":"9b0bfc38415de3bbdb2858d0bf6ed7e183f36528d2efaca79d6fb9d25ebd343c"} Oct 08 14:22:06 crc kubenswrapper[4789]: I1008 14:22:06.948893 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b9fe83-3af2-413e-b7fd-cabe53206482","Type":"ContainerDied","Data":"9d6eb57b02ece30ff5c215cbb5e23554bda0103123705a057368a3d0a1c0212c"} Oct 08 14:22:10 crc kubenswrapper[4789]: I1008 14:22:10.093939 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79cc7df9c4-dd4dv" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.162:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.162:8443: connect: connection refused" Oct 08 14:22:19 crc kubenswrapper[4789]: I1008 14:22:19.036449 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5c9c77979c-fl8sh" Oct 08 14:22:20 crc kubenswrapper[4789]: I1008 14:22:20.094722 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-79cc7df9c4-dd4dv" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.162:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.162:8443: connect: connection refused" Oct 08 14:22:20 crc kubenswrapper[4789]: I1008 14:22:20.095130 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.432016 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.433643 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.437407 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.437605 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.437762 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-xrvvb" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.443685 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.512760 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hlgm\" (UniqueName: \"kubernetes.io/projected/2ba79b7a-39cd-4f22-8094-672ef7b58e76-kube-api-access-2hlgm\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.512921 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2ba79b7a-39cd-4f22-8094-672ef7b58e76-openstack-config-secret\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.513025 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2ba79b7a-39cd-4f22-8094-672ef7b58e76-openstack-config\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.513100 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba79b7a-39cd-4f22-8094-672ef7b58e76-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.614440 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hlgm\" (UniqueName: \"kubernetes.io/projected/2ba79b7a-39cd-4f22-8094-672ef7b58e76-kube-api-access-2hlgm\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.614512 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2ba79b7a-39cd-4f22-8094-672ef7b58e76-openstack-config-secret\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.614564 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2ba79b7a-39cd-4f22-8094-672ef7b58e76-openstack-config\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.614592 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba79b7a-39cd-4f22-8094-672ef7b58e76-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.615612 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2ba79b7a-39cd-4f22-8094-672ef7b58e76-openstack-config\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.620848 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2ba79b7a-39cd-4f22-8094-672ef7b58e76-openstack-config-secret\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.623185 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba79b7a-39cd-4f22-8094-672ef7b58e76-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.643872 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hlgm\" (UniqueName: \"kubernetes.io/projected/2ba79b7a-39cd-4f22-8094-672ef7b58e76-kube-api-access-2hlgm\") pod \"openstackclient\" (UID: \"2ba79b7a-39cd-4f22-8094-672ef7b58e76\") " pod="openstack/openstackclient" Oct 08 14:22:24 crc kubenswrapper[4789]: I1008 14:22:24.757113 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.137483 4789 generic.go:334] "Generic (PLEG): container finished" podID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerID="b9f99165330ce21d34a9c1a6024be68840c5bbdb942c92c790f1bada38c7a935" exitCode=137 Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.137793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cc7df9c4-dd4dv" event={"ID":"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0","Type":"ContainerDied","Data":"b9f99165330ce21d34a9c1a6024be68840c5bbdb942c92c790f1bada38c7a935"} Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.282025 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.328236 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.432118 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f2mq\" (UniqueName: \"kubernetes.io/projected/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-kube-api-access-6f2mq\") pod \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.432172 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-config-data\") pod \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.432310 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-combined-ca-bundle\") pod \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.432348 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-logs\") pod \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.432426 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-secret-key\") pod \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.432458 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-tls-certs\") pod \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.432490 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-scripts\") pod \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\" (UID: \"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0\") " Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.432731 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-logs" (OuterVolumeSpecName: "logs") pod "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" (UID: "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.432975 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.450024 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" (UID: "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.450341 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-kube-api-access-6f2mq" (OuterVolumeSpecName: "kube-api-access-6f2mq") pod "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" (UID: "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0"). InnerVolumeSpecName "kube-api-access-6f2mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.463177 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-scripts" (OuterVolumeSpecName: "scripts") pod "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" (UID: "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.464931 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" (UID: "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.466598 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-config-data" (OuterVolumeSpecName: "config-data") pod "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" (UID: "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.495617 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" (UID: "cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.534566 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.534607 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.534618 4789 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.534629 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.534638 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f2mq\" (UniqueName: \"kubernetes.io/projected/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-kube-api-access-6f2mq\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:25 crc kubenswrapper[4789]: I1008 14:22:25.534648 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.151406 4789 generic.go:334] "Generic (PLEG): container finished" podID="ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" containerID="933f8f193cc92a17c8ff9a268982d0bc59d0534b608fb4c11e4b1f07def68aa2" exitCode=0 Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.151489 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-flz9d" event={"ID":"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9","Type":"ContainerDied","Data":"933f8f193cc92a17c8ff9a268982d0bc59d0534b608fb4c11e4b1f07def68aa2"} Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.153533 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2ba79b7a-39cd-4f22-8094-672ef7b58e76","Type":"ContainerStarted","Data":"c061265f258671cc8e6e84635883722c7b80dea6185d84e8f6845f7d7f085dac"} Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.155916 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79cc7df9c4-dd4dv" event={"ID":"cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0","Type":"ContainerDied","Data":"4d9c991bf33fcb9e0e5e500b76f0ae493c1edd08105bb669fbb7830a12775e13"} Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.155974 4789 scope.go:117] "RemoveContainer" containerID="3759762f30e8117b005cedf97abf2332be53ee3440a1616479c3c9aa5ed5b7c7" Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.156138 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79cc7df9c4-dd4dv" Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.201103 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79cc7df9c4-dd4dv"] Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.207549 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79cc7df9c4-dd4dv"] Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.348872 4789 scope.go:117] "RemoveContainer" containerID="b9f99165330ce21d34a9c1a6024be68840c5bbdb942c92c790f1bada38c7a935" Oct 08 14:22:26 crc kubenswrapper[4789]: I1008 14:22:26.742846 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" path="/var/lib/kubelet/pods/cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0/volumes" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.189471 4789 generic.go:334] "Generic (PLEG): container finished" podID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerID="2243dc665e6aa408ef5e8d4e4ede44e6f00de22af3a81a8d794d074e1b3df648" exitCode=1 Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.189504 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"33bbfc8e-64bd-480c-b755-dd802cba53fe","Type":"ContainerDied","Data":"2243dc665e6aa408ef5e8d4e4ede44e6f00de22af3a81a8d794d074e1b3df648"} Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.190620 4789 scope.go:117] "RemoveContainer" containerID="2243dc665e6aa408ef5e8d4e4ede44e6f00de22af3a81a8d794d074e1b3df648" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.495701 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-flz9d" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.680247 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-logs\") pod \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.680355 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-combined-ca-bundle\") pod \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.680490 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-scripts\") pod \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.680569 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkf94\" (UniqueName: \"kubernetes.io/projected/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-kube-api-access-nkf94\") pod \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.680741 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-config-data\") pod \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\" (UID: \"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9\") " Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.682385 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-logs" (OuterVolumeSpecName: "logs") pod "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" (UID: "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.702346 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-kube-api-access-nkf94" (OuterVolumeSpecName: "kube-api-access-nkf94") pod "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" (UID: "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9"). InnerVolumeSpecName "kube-api-access-nkf94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.706007 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-scripts" (OuterVolumeSpecName: "scripts") pod "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" (UID: "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.725227 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" (UID: "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.775207 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-config-data" (OuterVolumeSpecName: "config-data") pod "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" (UID: "ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.783434 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.783478 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.783489 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.783503 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:27 crc kubenswrapper[4789]: I1008 14:22:27.783513 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkf94\" (UniqueName: \"kubernetes.io/projected/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9-kube-api-access-nkf94\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.203404 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-flz9d" event={"ID":"ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9","Type":"ContainerDied","Data":"92e51d8741296d37282ec649137fa18b40ffd6eee17458c0bca3394cdad5a180"} Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.203704 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92e51d8741296d37282ec649137fa18b40ffd6eee17458c0bca3394cdad5a180" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.203441 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-flz9d" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.217835 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"33bbfc8e-64bd-480c-b755-dd802cba53fe","Type":"ContainerStarted","Data":"440e900070209cbbb2fed3c3255ae9692fdbb53508c99ab21fd0f7653dc31a03"} Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.329719 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6c99f9dffd-c2zmb"] Oct 08 14:22:28 crc kubenswrapper[4789]: E1008 14:22:28.338469 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon-log" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.338499 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon-log" Oct 08 14:22:28 crc kubenswrapper[4789]: E1008 14:22:28.338509 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" containerName="placement-db-sync" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.338515 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" containerName="placement-db-sync" Oct 08 14:22:28 crc kubenswrapper[4789]: E1008 14:22:28.338548 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.338556 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.338747 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.338763 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee3bbd5-b7fe-46ab-93a2-a4bcb82be1d0" containerName="horizon-log" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.338772 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" containerName="placement-db-sync" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.339742 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.342885 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.344178 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.344526 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-f8hg9" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.344686 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.344830 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.346950 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6c99f9dffd-c2zmb"] Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.498736 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-logs\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.498833 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnpdm\" (UniqueName: \"kubernetes.io/projected/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-kube-api-access-bnpdm\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.498863 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-internal-tls-certs\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.498886 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-scripts\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.498939 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-config-data\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.498955 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-public-tls-certs\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.499014 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-combined-ca-bundle\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.600785 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnpdm\" (UniqueName: \"kubernetes.io/projected/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-kube-api-access-bnpdm\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.600856 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-internal-tls-certs\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.600879 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-scripts\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.600947 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-config-data\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.600972 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-public-tls-certs\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.601496 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-combined-ca-bundle\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.601568 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-logs\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.602302 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-logs\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.609294 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-combined-ca-bundle\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.611173 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-internal-tls-certs\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.615357 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-public-tls-certs\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.620684 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-scripts\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.625199 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnpdm\" (UniqueName: \"kubernetes.io/projected/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-kube-api-access-bnpdm\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.630486 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c79d9b2e-fb14-4d73-a6f0-d8b0684001ac-config-data\") pod \"placement-6c99f9dffd-c2zmb\" (UID: \"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac\") " pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:28 crc kubenswrapper[4789]: I1008 14:22:28.680106 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.141065 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7bf8478fbc-fpmxn"] Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.144308 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.148398 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.148618 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.148745 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.157936 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7bf8478fbc-fpmxn"] Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.184845 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6c99f9dffd-c2zmb"] Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.236798 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c99f9dffd-c2zmb" event={"ID":"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac","Type":"ContainerStarted","Data":"2c4c0fba1147e1ac36a166424bd704d1577531b8da5ea66fb0bee1476ebc1174"} Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.315094 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba045b83-d6ec-4760-98dd-b330dca8ed99-run-httpd\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.315338 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba045b83-d6ec-4760-98dd-b330dca8ed99-log-httpd\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.315433 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-combined-ca-bundle\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.315566 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-internal-tls-certs\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.315798 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjq67\" (UniqueName: \"kubernetes.io/projected/ba045b83-d6ec-4760-98dd-b330dca8ed99-kube-api-access-cjq67\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.315895 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ba045b83-d6ec-4760-98dd-b330dca8ed99-etc-swift\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.316021 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-public-tls-certs\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.316128 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-config-data\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.417700 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ba045b83-d6ec-4760-98dd-b330dca8ed99-etc-swift\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.419347 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-public-tls-certs\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.419522 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-config-data\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.419714 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba045b83-d6ec-4760-98dd-b330dca8ed99-run-httpd\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.419810 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba045b83-d6ec-4760-98dd-b330dca8ed99-log-httpd\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.419907 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-combined-ca-bundle\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.420033 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-internal-tls-certs\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.420158 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjq67\" (UniqueName: \"kubernetes.io/projected/ba045b83-d6ec-4760-98dd-b330dca8ed99-kube-api-access-cjq67\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.420463 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba045b83-d6ec-4760-98dd-b330dca8ed99-run-httpd\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.420879 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba045b83-d6ec-4760-98dd-b330dca8ed99-log-httpd\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.422563 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-config-data\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.424934 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ba045b83-d6ec-4760-98dd-b330dca8ed99-etc-swift\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.425411 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-internal-tls-certs\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.425490 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-public-tls-certs\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.425972 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba045b83-d6ec-4760-98dd-b330dca8ed99-combined-ca-bundle\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.443952 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjq67\" (UniqueName: \"kubernetes.io/projected/ba045b83-d6ec-4760-98dd-b330dca8ed99-kube-api-access-cjq67\") pod \"swift-proxy-7bf8478fbc-fpmxn\" (UID: \"ba045b83-d6ec-4760-98dd-b330dca8ed99\") " pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.485919 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.796169 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 14:22:29 crc kubenswrapper[4789]: I1008 14:22:29.822783 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 08 14:22:30 crc kubenswrapper[4789]: I1008 14:22:30.259162 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c99f9dffd-c2zmb" event={"ID":"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac","Type":"ContainerStarted","Data":"76600be0c0f030a0244ba7ffa5bcbfc2801f976962d44945a2a1d17e9ccb2b55"} Oct 08 14:22:30 crc kubenswrapper[4789]: I1008 14:22:30.259505 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 14:22:30 crc kubenswrapper[4789]: I1008 14:22:30.290424 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 08 14:22:32 crc kubenswrapper[4789]: I1008 14:22:32.414213 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7bf8478fbc-fpmxn"] Oct 08 14:22:37 crc kubenswrapper[4789]: I1008 14:22:37.333802 4789 generic.go:334] "Generic (PLEG): container finished" podID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerID="7490defa0c96cd57eefcb7a504e9e31cf7452f614d3e942bcb01b991fcbe0ad8" exitCode=137 Oct 08 14:22:37 crc kubenswrapper[4789]: I1008 14:22:37.333925 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b9fe83-3af2-413e-b7fd-cabe53206482","Type":"ContainerDied","Data":"7490defa0c96cd57eefcb7a504e9e31cf7452f614d3e942bcb01b991fcbe0ad8"} Oct 08 14:22:40 crc kubenswrapper[4789]: I1008 14:22:40.975293 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.155:3000/\": dial tcp 10.217.0.155:3000: connect: connection refused" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.386347 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" event={"ID":"ba045b83-d6ec-4760-98dd-b330dca8ed99","Type":"ContainerStarted","Data":"af4002ea3839c294113234e7e91011ccd8061516f31ecea05acb0f691c8eaf16"} Oct 08 14:22:43 crc kubenswrapper[4789]: E1008 14:22:43.554856 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-openstackclient:watcher_latest" Oct 08 14:22:43 crc kubenswrapper[4789]: E1008 14:22:43.555259 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-openstackclient:watcher_latest" Oct 08 14:22:43 crc kubenswrapper[4789]: E1008 14:22:43.555464 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:38.102.83.222:5001/podified-master-centos10/openstack-openstackclient:watcher_latest,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf8h548h685h67ch89h7bh647h598hfh79h697hbh89h648h557h94h5ddh9ch576h695h544h68h5b8h599h78h57h5d7h5f8hdfh568h675h669q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_CA_CERT,Value:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2hlgm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(2ba79b7a-39cd-4f22-8094-672ef7b58e76): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:22:43 crc kubenswrapper[4789]: E1008 14:22:43.556662 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="2ba79b7a-39cd-4f22-8094-672ef7b58e76" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.742341 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.888479 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-log-httpd\") pod \"40b9fe83-3af2-413e-b7fd-cabe53206482\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.888630 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-config-data\") pod \"40b9fe83-3af2-413e-b7fd-cabe53206482\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.888710 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2czbg\" (UniqueName: \"kubernetes.io/projected/40b9fe83-3af2-413e-b7fd-cabe53206482-kube-api-access-2czbg\") pod \"40b9fe83-3af2-413e-b7fd-cabe53206482\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.888753 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-run-httpd\") pod \"40b9fe83-3af2-413e-b7fd-cabe53206482\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.888799 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-scripts\") pod \"40b9fe83-3af2-413e-b7fd-cabe53206482\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.888852 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-sg-core-conf-yaml\") pod \"40b9fe83-3af2-413e-b7fd-cabe53206482\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.888878 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-combined-ca-bundle\") pod \"40b9fe83-3af2-413e-b7fd-cabe53206482\" (UID: \"40b9fe83-3af2-413e-b7fd-cabe53206482\") " Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.889527 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "40b9fe83-3af2-413e-b7fd-cabe53206482" (UID: "40b9fe83-3af2-413e-b7fd-cabe53206482"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.889645 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "40b9fe83-3af2-413e-b7fd-cabe53206482" (UID: "40b9fe83-3af2-413e-b7fd-cabe53206482"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.890695 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.890904 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40b9fe83-3af2-413e-b7fd-cabe53206482-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.892637 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40b9fe83-3af2-413e-b7fd-cabe53206482-kube-api-access-2czbg" (OuterVolumeSpecName: "kube-api-access-2czbg") pod "40b9fe83-3af2-413e-b7fd-cabe53206482" (UID: "40b9fe83-3af2-413e-b7fd-cabe53206482"). InnerVolumeSpecName "kube-api-access-2czbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.894319 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-scripts" (OuterVolumeSpecName: "scripts") pod "40b9fe83-3af2-413e-b7fd-cabe53206482" (UID: "40b9fe83-3af2-413e-b7fd-cabe53206482"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.918493 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "40b9fe83-3af2-413e-b7fd-cabe53206482" (UID: "40b9fe83-3af2-413e-b7fd-cabe53206482"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.935720 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40b9fe83-3af2-413e-b7fd-cabe53206482" (UID: "40b9fe83-3af2-413e-b7fd-cabe53206482"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.983775 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-config-data" (OuterVolumeSpecName: "config-data") pod "40b9fe83-3af2-413e-b7fd-cabe53206482" (UID: "40b9fe83-3af2-413e-b7fd-cabe53206482"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.993452 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.993507 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2czbg\" (UniqueName: \"kubernetes.io/projected/40b9fe83-3af2-413e-b7fd-cabe53206482-kube-api-access-2czbg\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.993527 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.993549 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:43 crc kubenswrapper[4789]: I1008 14:22:43.993568 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40b9fe83-3af2-413e-b7fd-cabe53206482-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.398452 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c99f9dffd-c2zmb" event={"ID":"c79d9b2e-fb14-4d73-a6f0-d8b0684001ac","Type":"ContainerStarted","Data":"33950aba3bf4fe25200f550bd2a8c029b7e5dc4d6c0f9a4a768edbcce3f789c3"} Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.398870 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.403671 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" event={"ID":"ba045b83-d6ec-4760-98dd-b330dca8ed99","Type":"ContainerStarted","Data":"d8b48118ac514e3f6190954fffe7a1c4c93341a742c47e8bf214b7cfb7989f14"} Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.403730 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" event={"ID":"ba045b83-d6ec-4760-98dd-b330dca8ed99","Type":"ContainerStarted","Data":"1479b44e0564a5a38ad9a770c20761cec766c9e559427502f245389dc6e1f6e2"} Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.404361 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.404383 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.407642 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40b9fe83-3af2-413e-b7fd-cabe53206482","Type":"ContainerDied","Data":"7f53b0e2716ce580cd9090b6275b9fc4a6b1faf7748fe0596b7225d73e0a5a9e"} Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.407686 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.407700 4789 scope.go:117] "RemoveContainer" containerID="9b0bfc38415de3bbdb2858d0bf6ed7e183f36528d2efaca79d6fb9d25ebd343c" Oct 08 14:22:44 crc kubenswrapper[4789]: E1008 14:22:44.409511 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-openstackclient:watcher_latest\\\"\"" pod="openstack/openstackclient" podUID="2ba79b7a-39cd-4f22-8094-672ef7b58e76" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.433275 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6c99f9dffd-c2zmb" podStartSLOduration=16.433252208 podStartE2EDuration="16.433252208s" podCreationTimestamp="2025-10-08 14:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:22:44.424689516 +0000 UTC m=+1304.331437008" watchObservedRunningTime="2025-10-08 14:22:44.433252208 +0000 UTC m=+1304.339999710" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.450909 4789 scope.go:117] "RemoveContainer" containerID="9d6eb57b02ece30ff5c215cbb5e23554bda0103123705a057368a3d0a1c0212c" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.468795 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" podStartSLOduration=15.468776041 podStartE2EDuration="15.468776041s" podCreationTimestamp="2025-10-08 14:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:22:44.463337653 +0000 UTC m=+1304.370085165" watchObservedRunningTime="2025-10-08 14:22:44.468776041 +0000 UTC m=+1304.375523533" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.481218 4789 scope.go:117] "RemoveContainer" containerID="7490defa0c96cd57eefcb7a504e9e31cf7452f614d3e942bcb01b991fcbe0ad8" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.520931 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.541943 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.583429 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:22:44 crc kubenswrapper[4789]: E1008 14:22:44.583841 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="sg-core" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.583859 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="sg-core" Oct 08 14:22:44 crc kubenswrapper[4789]: E1008 14:22:44.583873 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="proxy-httpd" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.583880 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="proxy-httpd" Oct 08 14:22:44 crc kubenswrapper[4789]: E1008 14:22:44.583893 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="ceilometer-notification-agent" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.583899 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="ceilometer-notification-agent" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.584110 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="proxy-httpd" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.584120 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="ceilometer-notification-agent" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.584130 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" containerName="sg-core" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.585695 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.589652 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.594920 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.602902 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.705129 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.705192 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-log-httpd\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.705280 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-scripts\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.705351 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.705388 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb7hx\" (UniqueName: \"kubernetes.io/projected/835cb27d-9cce-484d-95bd-ee88abf5a2dc-kube-api-access-xb7hx\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.705426 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-run-httpd\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.705451 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-config-data\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.742865 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40b9fe83-3af2-413e-b7fd-cabe53206482" path="/var/lib/kubelet/pods/40b9fe83-3af2-413e-b7fd-cabe53206482/volumes" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.807391 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.807460 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb7hx\" (UniqueName: \"kubernetes.io/projected/835cb27d-9cce-484d-95bd-ee88abf5a2dc-kube-api-access-xb7hx\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.807507 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-run-httpd\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.807536 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-config-data\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.807585 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.807605 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-log-httpd\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.807686 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-scripts\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.808383 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-run-httpd\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.808485 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-log-httpd\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.813050 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-scripts\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.813307 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-config-data\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.813725 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.814570 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.832055 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb7hx\" (UniqueName: \"kubernetes.io/projected/835cb27d-9cce-484d-95bd-ee88abf5a2dc-kube-api-access-xb7hx\") pod \"ceilometer-0\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " pod="openstack/ceilometer-0" Oct 08 14:22:44 crc kubenswrapper[4789]: I1008 14:22:44.943150 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:22:45 crc kubenswrapper[4789]: I1008 14:22:45.374394 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:22:45 crc kubenswrapper[4789]: W1008 14:22:45.376352 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod835cb27d_9cce_484d_95bd_ee88abf5a2dc.slice/crio-6d8b42937a4b740b7857685b32e1387f28238f8bdf6ad79dfdf0cd9375163261 WatchSource:0}: Error finding container 6d8b42937a4b740b7857685b32e1387f28238f8bdf6ad79dfdf0cd9375163261: Status 404 returned error can't find the container with id 6d8b42937a4b740b7857685b32e1387f28238f8bdf6ad79dfdf0cd9375163261 Oct 08 14:22:45 crc kubenswrapper[4789]: I1008 14:22:45.418252 4789 generic.go:334] "Generic (PLEG): container finished" podID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerID="440e900070209cbbb2fed3c3255ae9692fdbb53508c99ab21fd0f7653dc31a03" exitCode=1 Oct 08 14:22:45 crc kubenswrapper[4789]: I1008 14:22:45.418314 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"33bbfc8e-64bd-480c-b755-dd802cba53fe","Type":"ContainerDied","Data":"440e900070209cbbb2fed3c3255ae9692fdbb53508c99ab21fd0f7653dc31a03"} Oct 08 14:22:45 crc kubenswrapper[4789]: I1008 14:22:45.418348 4789 scope.go:117] "RemoveContainer" containerID="2243dc665e6aa408ef5e8d4e4ede44e6f00de22af3a81a8d794d074e1b3df648" Oct 08 14:22:45 crc kubenswrapper[4789]: I1008 14:22:45.418919 4789 scope.go:117] "RemoveContainer" containerID="440e900070209cbbb2fed3c3255ae9692fdbb53508c99ab21fd0f7653dc31a03" Oct 08 14:22:45 crc kubenswrapper[4789]: E1008 14:22:45.419170 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(33bbfc8e-64bd-480c-b755-dd802cba53fe)\"" pod="openstack/watcher-decision-engine-0" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" Oct 08 14:22:45 crc kubenswrapper[4789]: I1008 14:22:45.420034 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerStarted","Data":"6d8b42937a4b740b7857685b32e1387f28238f8bdf6ad79dfdf0cd9375163261"} Oct 08 14:22:45 crc kubenswrapper[4789]: I1008 14:22:45.423741 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:46 crc kubenswrapper[4789]: I1008 14:22:46.436324 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerStarted","Data":"da4dcaa236c25382f0333cf0697d930eadaf476ef72aaa894d3d401749065420"} Oct 08 14:22:46 crc kubenswrapper[4789]: I1008 14:22:46.436623 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerStarted","Data":"f66f2fc2457dc225f1d54437ce02e187a00ef23ef69788900141659196068692"} Oct 08 14:22:46 crc kubenswrapper[4789]: I1008 14:22:46.437857 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:22:47 crc kubenswrapper[4789]: I1008 14:22:47.357395 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:22:47 crc kubenswrapper[4789]: I1008 14:22:47.448434 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerStarted","Data":"2a0ef0e58fcc4cb78ad4e12987bc81057e0bfe689a2a190455ee1e42efc2f5dd"} Oct 08 14:22:48 crc kubenswrapper[4789]: I1008 14:22:48.458969 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerStarted","Data":"d9686918aa16832170c166e4e73c2e45e5490b24d9f856084bdc0e6b31728181"} Oct 08 14:22:48 crc kubenswrapper[4789]: I1008 14:22:48.459351 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 14:22:48 crc kubenswrapper[4789]: I1008 14:22:48.488455 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.819271626 podStartE2EDuration="4.488434602s" podCreationTimestamp="2025-10-08 14:22:44 +0000 UTC" firstStartedPulling="2025-10-08 14:22:45.378670327 +0000 UTC m=+1305.285417819" lastFinishedPulling="2025-10-08 14:22:48.047833303 +0000 UTC m=+1307.954580795" observedRunningTime="2025-10-08 14:22:48.478539874 +0000 UTC m=+1308.385287386" watchObservedRunningTime="2025-10-08 14:22:48.488434602 +0000 UTC m=+1308.395182094" Oct 08 14:22:49 crc kubenswrapper[4789]: I1008 14:22:49.497300 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:49 crc kubenswrapper[4789]: I1008 14:22:49.498747 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7bf8478fbc-fpmxn" Oct 08 14:22:49 crc kubenswrapper[4789]: I1008 14:22:49.795804 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 14:22:49 crc kubenswrapper[4789]: I1008 14:22:49.796118 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 14:22:49 crc kubenswrapper[4789]: I1008 14:22:49.796222 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 14:22:49 crc kubenswrapper[4789]: I1008 14:22:49.796760 4789 scope.go:117] "RemoveContainer" containerID="440e900070209cbbb2fed3c3255ae9692fdbb53508c99ab21fd0f7653dc31a03" Oct 08 14:22:49 crc kubenswrapper[4789]: E1008 14:22:49.796993 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(33bbfc8e-64bd-480c-b755-dd802cba53fe)\"" pod="openstack/watcher-decision-engine-0" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" Oct 08 14:22:50 crc kubenswrapper[4789]: I1008 14:22:50.476761 4789 generic.go:334] "Generic (PLEG): container finished" podID="dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" containerID="c6a9a9f5178d5155c28bdee0e913b49dcfa8d0018b023832fff25943780f9225" exitCode=0 Oct 08 14:22:50 crc kubenswrapper[4789]: I1008 14:22:50.477666 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hks4r" event={"ID":"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0","Type":"ContainerDied","Data":"c6a9a9f5178d5155c28bdee0e913b49dcfa8d0018b023832fff25943780f9225"} Oct 08 14:22:50 crc kubenswrapper[4789]: I1008 14:22:50.478462 4789 scope.go:117] "RemoveContainer" containerID="440e900070209cbbb2fed3c3255ae9692fdbb53508c99ab21fd0f7653dc31a03" Oct 08 14:22:50 crc kubenswrapper[4789]: E1008 14:22:50.478652 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(33bbfc8e-64bd-480c-b755-dd802cba53fe)\"" pod="openstack/watcher-decision-engine-0" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" Oct 08 14:22:51 crc kubenswrapper[4789]: I1008 14:22:51.839421 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hks4r" Oct 08 14:22:51 crc kubenswrapper[4789]: I1008 14:22:51.948777 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-combined-ca-bundle\") pod \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " Oct 08 14:22:51 crc kubenswrapper[4789]: I1008 14:22:51.948896 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-db-sync-config-data\") pod \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " Oct 08 14:22:51 crc kubenswrapper[4789]: I1008 14:22:51.949196 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hzzj\" (UniqueName: \"kubernetes.io/projected/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-kube-api-access-4hzzj\") pod \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\" (UID: \"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0\") " Oct 08 14:22:51 crc kubenswrapper[4789]: I1008 14:22:51.954856 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" (UID: "dccf34c5-6e31-4d04-9478-d4e4edcc6ab0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:51 crc kubenswrapper[4789]: I1008 14:22:51.956139 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-kube-api-access-4hzzj" (OuterVolumeSpecName: "kube-api-access-4hzzj") pod "dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" (UID: "dccf34c5-6e31-4d04-9478-d4e4edcc6ab0"). InnerVolumeSpecName "kube-api-access-4hzzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:22:51 crc kubenswrapper[4789]: I1008 14:22:51.992193 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" (UID: "dccf34c5-6e31-4d04-9478-d4e4edcc6ab0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.050831 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hzzj\" (UniqueName: \"kubernetes.io/projected/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-kube-api-access-4hzzj\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.050866 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.050875 4789 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.495458 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hks4r" event={"ID":"dccf34c5-6e31-4d04-9478-d4e4edcc6ab0","Type":"ContainerDied","Data":"a1d6db17cd9aece796ec8bfbc227ee61b8f1157daff1c66794a9f4e56247fe34"} Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.495510 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1d6db17cd9aece796ec8bfbc227ee61b8f1157daff1c66794a9f4e56247fe34" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.495510 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hks4r" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.751173 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-56d947cc6f-sgvkl"] Oct 08 14:22:52 crc kubenswrapper[4789]: E1008 14:22:52.751815 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" containerName="barbican-db-sync" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.751837 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" containerName="barbican-db-sync" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.752105 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" containerName="barbican-db-sync" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.753446 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.760439 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.760686 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-svstf" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.760809 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.786705 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-56d947cc6f-sgvkl"] Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.803257 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5db9877796-wf2x5"] Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.804884 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.807526 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.856441 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5db9877796-wf2x5"] Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.870029 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-logs\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.870064 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-config-data-custom\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.870103 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-combined-ca-bundle\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.870144 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-config-data\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.870159 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzgr9\" (UniqueName: \"kubernetes.io/projected/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-kube-api-access-jzgr9\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.899613 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58c4748c7f-2gncr"] Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.902604 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.934900 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58c4748c7f-2gncr"] Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972311 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01d2c41c-c841-46c8-84a5-d03987977dba-config-data-custom\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972371 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-logs\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972391 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01d2c41c-c841-46c8-84a5-d03987977dba-config-data\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972408 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-config-data-custom\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972441 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-combined-ca-bundle\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972489 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-config-data\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972517 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzgr9\" (UniqueName: \"kubernetes.io/projected/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-kube-api-access-jzgr9\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972574 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d2c41c-c841-46c8-84a5-d03987977dba-combined-ca-bundle\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972612 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01d2c41c-c841-46c8-84a5-d03987977dba-logs\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.972714 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfbvs\" (UniqueName: \"kubernetes.io/projected/01d2c41c-c841-46c8-84a5-d03987977dba-kube-api-access-qfbvs\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.973156 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-logs\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.979189 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-combined-ca-bundle\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.979656 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-config-data-custom\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.998027 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-9cc5c8f44-4k7vr"] Oct 08 14:22:52 crc kubenswrapper[4789]: I1008 14:22:52.999200 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-config-data\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.000696 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.002089 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzgr9\" (UniqueName: \"kubernetes.io/projected/3028d0aa-cb65-4b14-aebb-7b9cb4910eed-kube-api-access-jzgr9\") pod \"barbican-worker-56d947cc6f-sgvkl\" (UID: \"3028d0aa-cb65-4b14-aebb-7b9cb4910eed\") " pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.005675 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.026666 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9cc5c8f44-4k7vr"] Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.085285 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01d2c41c-c841-46c8-84a5-d03987977dba-config-data\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091204 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcf6l\" (UniqueName: \"kubernetes.io/projected/0d3d0886-9352-4436-add7-40d813f0eeb1-kube-api-access-dcf6l\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091453 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-swift-storage-0\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091491 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d2c41c-c841-46c8-84a5-d03987977dba-combined-ca-bundle\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091572 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01d2c41c-c841-46c8-84a5-d03987977dba-logs\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091608 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-sb\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091758 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-config\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091820 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-svc\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091849 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-nb\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091925 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfbvs\" (UniqueName: \"kubernetes.io/projected/01d2c41c-c841-46c8-84a5-d03987977dba-kube-api-access-qfbvs\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.091965 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01d2c41c-c841-46c8-84a5-d03987977dba-config-data-custom\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.095699 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01d2c41c-c841-46c8-84a5-d03987977dba-config-data\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.096742 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01d2c41c-c841-46c8-84a5-d03987977dba-logs\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.099936 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d2c41c-c841-46c8-84a5-d03987977dba-combined-ca-bundle\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.105056 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-56d947cc6f-sgvkl" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.115707 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01d2c41c-c841-46c8-84a5-d03987977dba-config-data-custom\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.120568 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfbvs\" (UniqueName: \"kubernetes.io/projected/01d2c41c-c841-46c8-84a5-d03987977dba-kube-api-access-qfbvs\") pod \"barbican-keystone-listener-5db9877796-wf2x5\" (UID: \"01d2c41c-c841-46c8-84a5-d03987977dba\") " pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.126487 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193370 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-config\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193454 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-svc\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193477 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-nb\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193518 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data-custom\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193544 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193576 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-combined-ca-bundle\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193625 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcf6l\" (UniqueName: \"kubernetes.io/projected/0d3d0886-9352-4436-add7-40d813f0eeb1-kube-api-access-dcf6l\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193660 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/094807cf-ec15-4149-9967-5c52e35f93c2-logs\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193725 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-swift-storage-0\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193764 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prxrr\" (UniqueName: \"kubernetes.io/projected/094807cf-ec15-4149-9967-5c52e35f93c2-kube-api-access-prxrr\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.193793 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-sb\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.194886 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-sb\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.195623 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-config\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.196803 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-nb\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.196868 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-swift-storage-0\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.197196 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-svc\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.220610 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcf6l\" (UniqueName: \"kubernetes.io/projected/0d3d0886-9352-4436-add7-40d813f0eeb1-kube-api-access-dcf6l\") pod \"dnsmasq-dns-58c4748c7f-2gncr\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.241820 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.295383 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/094807cf-ec15-4149-9967-5c52e35f93c2-logs\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.295475 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prxrr\" (UniqueName: \"kubernetes.io/projected/094807cf-ec15-4149-9967-5c52e35f93c2-kube-api-access-prxrr\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.295546 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data-custom\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.295564 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.295589 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-combined-ca-bundle\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.297269 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/094807cf-ec15-4149-9967-5c52e35f93c2-logs\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.301641 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.302675 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-combined-ca-bundle\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.306816 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data-custom\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.318746 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prxrr\" (UniqueName: \"kubernetes.io/projected/094807cf-ec15-4149-9967-5c52e35f93c2-kube-api-access-prxrr\") pod \"barbican-api-9cc5c8f44-4k7vr\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.319882 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.534369 4789 generic.go:334] "Generic (PLEG): container finished" podID="ef2b1def-1464-4d5e-8d4e-44a4e505fa40" containerID="0b0da7215214f8a92093bc4106a2506ee926eac0474f267783e915ef17c69b45" exitCode=0 Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.534446 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tblv4" event={"ID":"ef2b1def-1464-4d5e-8d4e-44a4e505fa40","Type":"ContainerDied","Data":"0b0da7215214f8a92093bc4106a2506ee926eac0474f267783e915ef17c69b45"} Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.701033 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-56d947cc6f-sgvkl"] Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.848245 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5db9877796-wf2x5"] Oct 08 14:22:53 crc kubenswrapper[4789]: I1008 14:22:53.964268 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58c4748c7f-2gncr"] Oct 08 14:22:53 crc kubenswrapper[4789]: W1008 14:22:53.972098 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d3d0886_9352_4436_add7_40d813f0eeb1.slice/crio-8c1ec34d5f847a17a30543eb01fe695710e4907c10b19cb24dc43500fa7ff323 WatchSource:0}: Error finding container 8c1ec34d5f847a17a30543eb01fe695710e4907c10b19cb24dc43500fa7ff323: Status 404 returned error can't find the container with id 8c1ec34d5f847a17a30543eb01fe695710e4907c10b19cb24dc43500fa7ff323 Oct 08 14:22:54 crc kubenswrapper[4789]: I1008 14:22:54.007011 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9cc5c8f44-4k7vr"] Oct 08 14:22:54 crc kubenswrapper[4789]: I1008 14:22:54.554388 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" event={"ID":"01d2c41c-c841-46c8-84a5-d03987977dba","Type":"ContainerStarted","Data":"ae3b6a604d877e5e9e6b23469d90c74742c1531b94c367fff4c583779044f591"} Oct 08 14:22:54 crc kubenswrapper[4789]: I1008 14:22:54.556360 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-56d947cc6f-sgvkl" event={"ID":"3028d0aa-cb65-4b14-aebb-7b9cb4910eed","Type":"ContainerStarted","Data":"2d654596319113316d41ce51e94dfb8594b4a7a1804c3c7efbd08817fefd98fd"} Oct 08 14:22:54 crc kubenswrapper[4789]: I1008 14:22:54.559024 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" event={"ID":"0d3d0886-9352-4436-add7-40d813f0eeb1","Type":"ContainerStarted","Data":"8c1ec34d5f847a17a30543eb01fe695710e4907c10b19cb24dc43500fa7ff323"} Oct 08 14:22:54 crc kubenswrapper[4789]: I1008 14:22:54.561032 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9cc5c8f44-4k7vr" event={"ID":"094807cf-ec15-4149-9967-5c52e35f93c2","Type":"ContainerStarted","Data":"5197cad88885c384fd6f2daddd3438d9a12fd275145e51ed550c04fe297700ef"} Oct 08 14:22:54 crc kubenswrapper[4789]: I1008 14:22:54.561077 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9cc5c8f44-4k7vr" event={"ID":"094807cf-ec15-4149-9967-5c52e35f93c2","Type":"ContainerStarted","Data":"e73d55e2083d59c0e25e7ce5d100e7e3c529f52c60511d8ee591f62638671ace"} Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.300295 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tblv4" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.330119 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.331472 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="proxy-httpd" containerID="cri-o://d9686918aa16832170c166e4e73c2e45e5490b24d9f856084bdc0e6b31728181" gracePeriod=30 Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.331622 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="sg-core" containerID="cri-o://2a0ef0e58fcc4cb78ad4e12987bc81057e0bfe689a2a190455ee1e42efc2f5dd" gracePeriod=30 Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.331669 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="ceilometer-notification-agent" containerID="cri-o://da4dcaa236c25382f0333cf0697d930eadaf476ef72aaa894d3d401749065420" gracePeriod=30 Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.331422 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="ceilometer-central-agent" containerID="cri-o://f66f2fc2457dc225f1d54437ce02e187a00ef23ef69788900141659196068692" gracePeriod=30 Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.447684 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-scripts\") pod \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.447813 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-etc-machine-id\") pod \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.447837 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-combined-ca-bundle\") pod \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.447864 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-config-data\") pod \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.447913 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs6qq\" (UniqueName: \"kubernetes.io/projected/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-kube-api-access-bs6qq\") pod \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.448970 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-db-sync-config-data\") pod \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\" (UID: \"ef2b1def-1464-4d5e-8d4e-44a4e505fa40\") " Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.451084 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ef2b1def-1464-4d5e-8d4e-44a4e505fa40" (UID: "ef2b1def-1464-4d5e-8d4e-44a4e505fa40"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.458106 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-scripts" (OuterVolumeSpecName: "scripts") pod "ef2b1def-1464-4d5e-8d4e-44a4e505fa40" (UID: "ef2b1def-1464-4d5e-8d4e-44a4e505fa40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.461883 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ef2b1def-1464-4d5e-8d4e-44a4e505fa40" (UID: "ef2b1def-1464-4d5e-8d4e-44a4e505fa40"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.464582 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-kube-api-access-bs6qq" (OuterVolumeSpecName: "kube-api-access-bs6qq") pod "ef2b1def-1464-4d5e-8d4e-44a4e505fa40" (UID: "ef2b1def-1464-4d5e-8d4e-44a4e505fa40"). InnerVolumeSpecName "kube-api-access-bs6qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.488466 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef2b1def-1464-4d5e-8d4e-44a4e505fa40" (UID: "ef2b1def-1464-4d5e-8d4e-44a4e505fa40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.551304 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.551338 4789 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.551352 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.551364 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs6qq\" (UniqueName: \"kubernetes.io/projected/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-kube-api-access-bs6qq\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.551400 4789 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.561110 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-config-data" (OuterVolumeSpecName: "config-data") pod "ef2b1def-1464-4d5e-8d4e-44a4e505fa40" (UID: "ef2b1def-1464-4d5e-8d4e-44a4e505fa40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.572401 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d3d0886-9352-4436-add7-40d813f0eeb1" containerID="8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498" exitCode=0 Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.573509 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" event={"ID":"0d3d0886-9352-4436-add7-40d813f0eeb1","Type":"ContainerDied","Data":"8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498"} Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.581236 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9cc5c8f44-4k7vr" event={"ID":"094807cf-ec15-4149-9967-5c52e35f93c2","Type":"ContainerStarted","Data":"e98ee6b5a44a0d4b42e2e77219fdfff8c7fd27d9c3e3e41e6edb6671fba0632e"} Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.582304 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.582333 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.585143 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tblv4" event={"ID":"ef2b1def-1464-4d5e-8d4e-44a4e505fa40","Type":"ContainerDied","Data":"e80fc9727fd4f143a032e21ac09accab896a79494a8ede322ee1fa9ffd51da8c"} Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.585198 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e80fc9727fd4f143a032e21ac09accab896a79494a8ede322ee1fa9ffd51da8c" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.585253 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tblv4" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.609746 4789 generic.go:334] "Generic (PLEG): container finished" podID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerID="d9686918aa16832170c166e4e73c2e45e5490b24d9f856084bdc0e6b31728181" exitCode=0 Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.609773 4789 generic.go:334] "Generic (PLEG): container finished" podID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerID="2a0ef0e58fcc4cb78ad4e12987bc81057e0bfe689a2a190455ee1e42efc2f5dd" exitCode=2 Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.609795 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerDied","Data":"d9686918aa16832170c166e4e73c2e45e5490b24d9f856084bdc0e6b31728181"} Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.609819 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerDied","Data":"2a0ef0e58fcc4cb78ad4e12987bc81057e0bfe689a2a190455ee1e42efc2f5dd"} Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.631870 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-9cc5c8f44-4k7vr" podStartSLOduration=3.631848337 podStartE2EDuration="3.631848337s" podCreationTimestamp="2025-10-08 14:22:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:22:55.623365877 +0000 UTC m=+1315.530113369" watchObservedRunningTime="2025-10-08 14:22:55.631848337 +0000 UTC m=+1315.538595829" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.652616 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef2b1def-1464-4d5e-8d4e-44a4e505fa40-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.852097 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 14:22:55 crc kubenswrapper[4789]: E1008 14:22:55.852527 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2b1def-1464-4d5e-8d4e-44a4e505fa40" containerName="cinder-db-sync" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.852539 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2b1def-1464-4d5e-8d4e-44a4e505fa40" containerName="cinder-db-sync" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.852725 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2b1def-1464-4d5e-8d4e-44a4e505fa40" containerName="cinder-db-sync" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.853740 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.862478 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.862478 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.866457 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-qw9dn" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.866665 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.879600 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.882726 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.882794 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67cd2\" (UniqueName: \"kubernetes.io/projected/c611766d-13b9-4814-bff6-0932c986de51-kube-api-access-67cd2\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.882883 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c611766d-13b9-4814-bff6-0932c986de51-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.882950 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-scripts\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.883019 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.883092 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.990942 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67cd2\" (UniqueName: \"kubernetes.io/projected/c611766d-13b9-4814-bff6-0932c986de51-kube-api-access-67cd2\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.991447 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c611766d-13b9-4814-bff6-0932c986de51-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.991504 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-scripts\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.991557 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.991625 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.991682 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.993198 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c611766d-13b9-4814-bff6-0932c986de51-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:55 crc kubenswrapper[4789]: I1008 14:22:55.997469 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.009137 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.010814 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58c4748c7f-2gncr"] Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.014218 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.016681 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-scripts\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.040672 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67cd2\" (UniqueName: \"kubernetes.io/projected/c611766d-13b9-4814-bff6-0932c986de51-kube-api-access-67cd2\") pod \"cinder-scheduler-0\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " pod="openstack/cinder-scheduler-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.042351 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f9bf94f7c-84pd2"] Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.044397 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.058706 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f9bf94f7c-84pd2"] Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.170027 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.171796 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.176869 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.185344 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.195870 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-config\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.195910 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.195946 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.195979 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-swift-storage-0\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.196029 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsdst\" (UniqueName: \"kubernetes.io/projected/632b4bda-256d-4237-aade-b2cdffbf2d93-kube-api-access-tsdst\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.196289 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-svc\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.205143 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297692 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297738 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-scripts\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297766 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-config\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297797 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297816 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297835 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data-custom\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297861 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-swift-storage-0\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297886 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb5kk\" (UniqueName: \"kubernetes.io/projected/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-kube-api-access-rb5kk\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297902 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-logs\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297929 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsdst\" (UniqueName: \"kubernetes.io/projected/632b4bda-256d-4237-aade-b2cdffbf2d93-kube-api-access-tsdst\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297943 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-svc\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.297969 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.298009 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.298976 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-config\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.299122 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.300029 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-svc\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.300159 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.302023 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-swift-storage-0\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.324734 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsdst\" (UniqueName: \"kubernetes.io/projected/632b4bda-256d-4237-aade-b2cdffbf2d93-kube-api-access-tsdst\") pod \"dnsmasq-dns-6f9bf94f7c-84pd2\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.399404 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.399466 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-scripts\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.399563 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data-custom\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.399620 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb5kk\" (UniqueName: \"kubernetes.io/projected/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-kube-api-access-rb5kk\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.399644 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-logs\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.399735 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.399783 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.405333 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-logs\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.406325 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.408160 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.408677 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-scripts\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.412894 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-59fddc7b48-jvk6w"] Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.414655 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data-custom\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.433425 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.434159 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-59fddc7b48-jvk6w"] Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.434252 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.435233 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.435277 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.438623 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.438933 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.440334 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb5kk\" (UniqueName: \"kubernetes.io/projected/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-kube-api-access-rb5kk\") pod \"cinder-api-0\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.582210 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.594628 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.605009 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-combined-ca-bundle\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.605088 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-public-tls-certs\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.605116 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-config-data\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.605162 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-internal-tls-certs\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.605185 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2vkb\" (UniqueName: \"kubernetes.io/projected/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-kube-api-access-l2vkb\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.605214 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-config-data-custom\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.605255 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-logs\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.694294 4789 generic.go:334] "Generic (PLEG): container finished" podID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerID="f66f2fc2457dc225f1d54437ce02e187a00ef23ef69788900141659196068692" exitCode=0 Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.694605 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerDied","Data":"f66f2fc2457dc225f1d54437ce02e187a00ef23ef69788900141659196068692"} Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.707288 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-public-tls-certs\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.707358 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-config-data\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.707449 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-internal-tls-certs\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.709951 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2vkb\" (UniqueName: \"kubernetes.io/projected/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-kube-api-access-l2vkb\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.710028 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-config-data-custom\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.710416 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-logs\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.711125 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-combined-ca-bundle\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.712846 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-internal-tls-certs\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.713159 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-logs\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.714012 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-config-data-custom\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.714217 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-public-tls-certs\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.725399 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-config-data\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.735647 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-combined-ca-bundle\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.745924 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2vkb\" (UniqueName: \"kubernetes.io/projected/1033617f-6aaf-4d50-bd32-fb0eaf6c99b8-kube-api-access-l2vkb\") pod \"barbican-api-59fddc7b48-jvk6w\" (UID: \"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8\") " pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:56 crc kubenswrapper[4789]: I1008 14:22:56.808405 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.001430 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.197932 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f9bf94f7c-84pd2"] Oct 08 14:22:57 crc kubenswrapper[4789]: W1008 14:22:57.275059 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod632b4bda_256d_4237_aade_b2cdffbf2d93.slice/crio-ab1c0ce8619a6e3796813ec892fda7994e5ae4c54d30dc4d000ce95030eadcb4 WatchSource:0}: Error finding container ab1c0ce8619a6e3796813ec892fda7994e5ae4c54d30dc4d000ce95030eadcb4: Status 404 returned error can't find the container with id ab1c0ce8619a6e3796813ec892fda7994e5ae4c54d30dc4d000ce95030eadcb4 Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.324364 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.472756 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-59fddc7b48-jvk6w"] Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.754882 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"057a92ec-1b02-446f-bc0e-3ac7d2344d2c","Type":"ContainerStarted","Data":"3c780b9ae041f2f359253c6120ff30cdc55859fbdf89df434c9389ae35d1d30b"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.760573 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-56d947cc6f-sgvkl" event={"ID":"3028d0aa-cb65-4b14-aebb-7b9cb4910eed","Type":"ContainerStarted","Data":"733bc07a07d353506b5bcf1f3347196790aba086ba7ce79cce762967c46f00bd"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.760628 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-56d947cc6f-sgvkl" event={"ID":"3028d0aa-cb65-4b14-aebb-7b9cb4910eed","Type":"ContainerStarted","Data":"1583a66a97720cec0f9daf206429a588875535d4dca7195bf6da0d53ce6fbc17"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.774502 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" event={"ID":"0d3d0886-9352-4436-add7-40d813f0eeb1","Type":"ContainerStarted","Data":"25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.774667 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" podUID="0d3d0886-9352-4436-add7-40d813f0eeb1" containerName="dnsmasq-dns" containerID="cri-o://25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5" gracePeriod=10 Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.775052 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.783444 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-56d947cc6f-sgvkl" podStartSLOduration=3.2287344940000002 podStartE2EDuration="5.783416008s" podCreationTimestamp="2025-10-08 14:22:52 +0000 UTC" firstStartedPulling="2025-10-08 14:22:53.692840136 +0000 UTC m=+1313.599587628" lastFinishedPulling="2025-10-08 14:22:56.24752165 +0000 UTC m=+1316.154269142" observedRunningTime="2025-10-08 14:22:57.78276673 +0000 UTC m=+1317.689514222" watchObservedRunningTime="2025-10-08 14:22:57.783416008 +0000 UTC m=+1317.690163510" Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.786191 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c611766d-13b9-4814-bff6-0932c986de51","Type":"ContainerStarted","Data":"77952a8d887f5c6fe9956b93f334f2f9b9bcf140a6d2521e1573f72b07e5f319"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.802259 4789 generic.go:334] "Generic (PLEG): container finished" podID="632b4bda-256d-4237-aade-b2cdffbf2d93" containerID="1efbe07b43550f1d5907871e1eefe02ab9ed15279288204ed952b44c292548ef" exitCode=0 Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.802356 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" event={"ID":"632b4bda-256d-4237-aade-b2cdffbf2d93","Type":"ContainerDied","Data":"1efbe07b43550f1d5907871e1eefe02ab9ed15279288204ed952b44c292548ef"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.802383 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" event={"ID":"632b4bda-256d-4237-aade-b2cdffbf2d93","Type":"ContainerStarted","Data":"ab1c0ce8619a6e3796813ec892fda7994e5ae4c54d30dc4d000ce95030eadcb4"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.832569 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" podStartSLOduration=5.832543969 podStartE2EDuration="5.832543969s" podCreationTimestamp="2025-10-08 14:22:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:22:57.819686341 +0000 UTC m=+1317.726433833" watchObservedRunningTime="2025-10-08 14:22:57.832543969 +0000 UTC m=+1317.739291471" Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.847479 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2ba79b7a-39cd-4f22-8094-672ef7b58e76","Type":"ContainerStarted","Data":"d902ecb0d998268191dd1a33346f56d69b63f96d34f8b79897c1a4213e2a6b8d"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.858691 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" event={"ID":"01d2c41c-c841-46c8-84a5-d03987977dba","Type":"ContainerStarted","Data":"aa6ac915b2341baf57f756e15f12de6e14d1430d13f4b14177013b09e91b372c"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.858738 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" event={"ID":"01d2c41c-c841-46c8-84a5-d03987977dba","Type":"ContainerStarted","Data":"bb5285e6b116bad0f490801460c2a3195646ee2ed02a060d9bd6832ad3195390"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.860291 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59fddc7b48-jvk6w" event={"ID":"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8","Type":"ContainerStarted","Data":"35a086322e88cb4cadc8a2222f00fc52a52c87f822aacd67a92f18c485347c20"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.860329 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59fddc7b48-jvk6w" event={"ID":"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8","Type":"ContainerStarted","Data":"564ba2ff933676e0fd598f452399829b0702131495ed81502bed070cd22f4a64"} Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.891699 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.810117103 podStartE2EDuration="33.891675732s" podCreationTimestamp="2025-10-08 14:22:24 +0000 UTC" firstStartedPulling="2025-10-08 14:22:25.2873361 +0000 UTC m=+1285.194083582" lastFinishedPulling="2025-10-08 14:22:56.368894719 +0000 UTC m=+1316.275642211" observedRunningTime="2025-10-08 14:22:57.872106191 +0000 UTC m=+1317.778853693" watchObservedRunningTime="2025-10-08 14:22:57.891675732 +0000 UTC m=+1317.798423224" Oct 08 14:22:57 crc kubenswrapper[4789]: I1008 14:22:57.987898 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5db9877796-wf2x5" podStartSLOduration=3.589689464 podStartE2EDuration="5.987840627s" podCreationTimestamp="2025-10-08 14:22:52 +0000 UTC" firstStartedPulling="2025-10-08 14:22:53.847644541 +0000 UTC m=+1313.754392033" lastFinishedPulling="2025-10-08 14:22:56.245795694 +0000 UTC m=+1316.152543196" observedRunningTime="2025-10-08 14:22:57.904466338 +0000 UTC m=+1317.811213830" watchObservedRunningTime="2025-10-08 14:22:57.987840627 +0000 UTC m=+1317.894588119" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.554679 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.670627 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-svc\") pod \"0d3d0886-9352-4436-add7-40d813f0eeb1\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.670715 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-sb\") pod \"0d3d0886-9352-4436-add7-40d813f0eeb1\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.670777 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-swift-storage-0\") pod \"0d3d0886-9352-4436-add7-40d813f0eeb1\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.670830 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-config\") pod \"0d3d0886-9352-4436-add7-40d813f0eeb1\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.670891 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-nb\") pod \"0d3d0886-9352-4436-add7-40d813f0eeb1\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.670941 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcf6l\" (UniqueName: \"kubernetes.io/projected/0d3d0886-9352-4436-add7-40d813f0eeb1-kube-api-access-dcf6l\") pod \"0d3d0886-9352-4436-add7-40d813f0eeb1\" (UID: \"0d3d0886-9352-4436-add7-40d813f0eeb1\") " Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.706172 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d3d0886-9352-4436-add7-40d813f0eeb1-kube-api-access-dcf6l" (OuterVolumeSpecName: "kube-api-access-dcf6l") pod "0d3d0886-9352-4436-add7-40d813f0eeb1" (UID: "0d3d0886-9352-4436-add7-40d813f0eeb1"). InnerVolumeSpecName "kube-api-access-dcf6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.772879 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcf6l\" (UniqueName: \"kubernetes.io/projected/0d3d0886-9352-4436-add7-40d813f0eeb1-kube-api-access-dcf6l\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.843229 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0d3d0886-9352-4436-add7-40d813f0eeb1" (UID: "0d3d0886-9352-4436-add7-40d813f0eeb1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.844408 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0d3d0886-9352-4436-add7-40d813f0eeb1" (UID: "0d3d0886-9352-4436-add7-40d813f0eeb1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.855056 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0d3d0886-9352-4436-add7-40d813f0eeb1" (UID: "0d3d0886-9352-4436-add7-40d813f0eeb1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.877294 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.877329 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.877339 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.893574 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d3d0886-9352-4436-add7-40d813f0eeb1" (UID: "0d3d0886-9352-4436-add7-40d813f0eeb1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.893604 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-config" (OuterVolumeSpecName: "config") pod "0d3d0886-9352-4436-add7-40d813f0eeb1" (UID: "0d3d0886-9352-4436-add7-40d813f0eeb1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.908114 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-59fddc7b48-jvk6w" event={"ID":"1033617f-6aaf-4d50-bd32-fb0eaf6c99b8","Type":"ContainerStarted","Data":"d10df8f6614961c2ef2e63590d2b508e1681d48b417ea8693c7928a05270af35"} Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.908310 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.931212 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"057a92ec-1b02-446f-bc0e-3ac7d2344d2c","Type":"ContainerStarted","Data":"e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49"} Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.937842 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-59fddc7b48-jvk6w" podStartSLOduration=2.937821939 podStartE2EDuration="2.937821939s" podCreationTimestamp="2025-10-08 14:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:22:58.933570364 +0000 UTC m=+1318.840317856" watchObservedRunningTime="2025-10-08 14:22:58.937821939 +0000 UTC m=+1318.844569431" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.941073 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d3d0886-9352-4436-add7-40d813f0eeb1" containerID="25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5" exitCode=0 Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.941139 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" event={"ID":"0d3d0886-9352-4436-add7-40d813f0eeb1","Type":"ContainerDied","Data":"25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5"} Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.941167 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" event={"ID":"0d3d0886-9352-4436-add7-40d813f0eeb1","Type":"ContainerDied","Data":"8c1ec34d5f847a17a30543eb01fe695710e4907c10b19cb24dc43500fa7ff323"} Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.941183 4789 scope.go:117] "RemoveContainer" containerID="25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.941311 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c4748c7f-2gncr" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.953699 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c611766d-13b9-4814-bff6-0932c986de51","Type":"ContainerStarted","Data":"d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba"} Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.964692 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" event={"ID":"632b4bda-256d-4237-aade-b2cdffbf2d93","Type":"ContainerStarted","Data":"cd96ed81cb45e4420f241225df836cd12fce3b507ef2cf97afc720a3e2df01a7"} Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.965240 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.974478 4789 scope.go:117] "RemoveContainer" containerID="8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.979086 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.979111 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d3d0886-9352-4436-add7-40d813f0eeb1-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:22:58 crc kubenswrapper[4789]: I1008 14:22:58.997449 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" podStartSLOduration=3.9974323050000002 podStartE2EDuration="3.997432305s" podCreationTimestamp="2025-10-08 14:22:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:22:58.987886926 +0000 UTC m=+1318.894634418" watchObservedRunningTime="2025-10-08 14:22:58.997432305 +0000 UTC m=+1318.904179797" Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.101248 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58c4748c7f-2gncr"] Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.111604 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58c4748c7f-2gncr"] Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.124355 4789 scope.go:117] "RemoveContainer" containerID="25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5" Oct 08 14:22:59 crc kubenswrapper[4789]: E1008 14:22:59.131692 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5\": container with ID starting with 25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5 not found: ID does not exist" containerID="25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5" Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.131735 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5"} err="failed to get container status \"25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5\": rpc error: code = NotFound desc = could not find container \"25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5\": container with ID starting with 25fbf708cb75d3c9eded8a5841c63461c1c07251ff11452e4e92436bd23965d5 not found: ID does not exist" Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.131781 4789 scope.go:117] "RemoveContainer" containerID="8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498" Oct 08 14:22:59 crc kubenswrapper[4789]: E1008 14:22:59.132387 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498\": container with ID starting with 8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498 not found: ID does not exist" containerID="8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498" Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.132417 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498"} err="failed to get container status \"8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498\": rpc error: code = NotFound desc = could not find container \"8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498\": container with ID starting with 8ce1ef77f94c1cdc6b212e0cb5d7b75cf551aa91f5605db213365e11e7894498 not found: ID does not exist" Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.701210 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.976229 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c611766d-13b9-4814-bff6-0932c986de51","Type":"ContainerStarted","Data":"142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179"} Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.977824 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"057a92ec-1b02-446f-bc0e-3ac7d2344d2c","Type":"ContainerStarted","Data":"3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11"} Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.977963 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.978102 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.978117 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerName="cinder-api-log" containerID="cri-o://e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49" gracePeriod=30 Oct 08 14:22:59 crc kubenswrapper[4789]: I1008 14:22:59.978146 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerName="cinder-api" containerID="cri-o://3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11" gracePeriod=30 Oct 08 14:23:00 crc kubenswrapper[4789]: I1008 14:23:00.002073 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.654945211 podStartE2EDuration="5.002055657s" podCreationTimestamp="2025-10-08 14:22:55 +0000 UTC" firstStartedPulling="2025-10-08 14:22:57.050512449 +0000 UTC m=+1316.957259941" lastFinishedPulling="2025-10-08 14:22:57.397622895 +0000 UTC m=+1317.304370387" observedRunningTime="2025-10-08 14:22:59.994132943 +0000 UTC m=+1319.900880455" watchObservedRunningTime="2025-10-08 14:23:00.002055657 +0000 UTC m=+1319.908803149" Oct 08 14:23:00 crc kubenswrapper[4789]: I1008 14:23:00.023089 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.023065717 podStartE2EDuration="4.023065717s" podCreationTimestamp="2025-10-08 14:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:00.020155418 +0000 UTC m=+1319.926902930" watchObservedRunningTime="2025-10-08 14:23:00.023065717 +0000 UTC m=+1319.929813229" Oct 08 14:23:00 crc kubenswrapper[4789]: I1008 14:23:00.744675 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d3d0886-9352-4436-add7-40d813f0eeb1" path="/var/lib/kubelet/pods/0d3d0886-9352-4436-add7-40d813f0eeb1/volumes" Oct 08 14:23:00 crc kubenswrapper[4789]: I1008 14:23:00.988453 4789 generic.go:334] "Generic (PLEG): container finished" podID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerID="e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49" exitCode=143 Oct 08 14:23:00 crc kubenswrapper[4789]: I1008 14:23:00.988510 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"057a92ec-1b02-446f-bc0e-3ac7d2344d2c","Type":"ContainerDied","Data":"e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49"} Oct 08 14:23:01 crc kubenswrapper[4789]: I1008 14:23:01.186159 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 14:23:02 crc kubenswrapper[4789]: I1008 14:23:02.330658 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6c99f9dffd-c2zmb" Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.730908 4789 scope.go:117] "RemoveContainer" containerID="440e900070209cbbb2fed3c3255ae9692fdbb53508c99ab21fd0f7653dc31a03" Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.802947 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-xll5s"] Oct 08 14:23:04 crc kubenswrapper[4789]: E1008 14:23:04.803728 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3d0886-9352-4436-add7-40d813f0eeb1" containerName="init" Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.803752 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3d0886-9352-4436-add7-40d813f0eeb1" containerName="init" Oct 08 14:23:04 crc kubenswrapper[4789]: E1008 14:23:04.803804 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3d0886-9352-4436-add7-40d813f0eeb1" containerName="dnsmasq-dns" Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.803812 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3d0886-9352-4436-add7-40d813f0eeb1" containerName="dnsmasq-dns" Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.804046 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d3d0886-9352-4436-add7-40d813f0eeb1" containerName="dnsmasq-dns" Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.804657 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xll5s" Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.820970 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xll5s"] Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.908194 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-2wfw9"] Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.910060 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2wfw9" Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.923693 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq88t\" (UniqueName: \"kubernetes.io/projected/150d7f22-fe1e-47cf-bf5e-7f97473446b8-kube-api-access-lq88t\") pod \"nova-api-db-create-xll5s\" (UID: \"150d7f22-fe1e-47cf-bf5e-7f97473446b8\") " pod="openstack/nova-api-db-create-xll5s" Oct 08 14:23:04 crc kubenswrapper[4789]: I1008 14:23:04.937260 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-2wfw9"] Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.013855 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.030115 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfsw4\" (UniqueName: \"kubernetes.io/projected/05a2fc22-aea2-42aa-810a-98bbf5ba7c98-kube-api-access-kfsw4\") pod \"nova-cell0-db-create-2wfw9\" (UID: \"05a2fc22-aea2-42aa-810a-98bbf5ba7c98\") " pod="openstack/nova-cell0-db-create-2wfw9" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.030331 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq88t\" (UniqueName: \"kubernetes.io/projected/150d7f22-fe1e-47cf-bf5e-7f97473446b8-kube-api-access-lq88t\") pod \"nova-api-db-create-xll5s\" (UID: \"150d7f22-fe1e-47cf-bf5e-7f97473446b8\") " pod="openstack/nova-api-db-create-xll5s" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.035569 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"33bbfc8e-64bd-480c-b755-dd802cba53fe","Type":"ContainerStarted","Data":"bbaf9b2938fb9a5809414790ba8a305f1118bf8d38764891a274c4d9bd99e4af"} Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.058086 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq88t\" (UniqueName: \"kubernetes.io/projected/150d7f22-fe1e-47cf-bf5e-7f97473446b8-kube-api-access-lq88t\") pod \"nova-api-db-create-xll5s\" (UID: \"150d7f22-fe1e-47cf-bf5e-7f97473446b8\") " pod="openstack/nova-api-db-create-xll5s" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.079039 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.098662 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-69jjr"] Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.100049 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-69jjr" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.124660 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-69jjr"] Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.131834 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfsw4\" (UniqueName: \"kubernetes.io/projected/05a2fc22-aea2-42aa-810a-98bbf5ba7c98-kube-api-access-kfsw4\") pod \"nova-cell0-db-create-2wfw9\" (UID: \"05a2fc22-aea2-42aa-810a-98bbf5ba7c98\") " pod="openstack/nova-cell0-db-create-2wfw9" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.183100 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfsw4\" (UniqueName: \"kubernetes.io/projected/05a2fc22-aea2-42aa-810a-98bbf5ba7c98-kube-api-access-kfsw4\") pod \"nova-cell0-db-create-2wfw9\" (UID: \"05a2fc22-aea2-42aa-810a-98bbf5ba7c98\") " pod="openstack/nova-cell0-db-create-2wfw9" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.208531 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xll5s" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.229511 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2wfw9" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.235495 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pztz\" (UniqueName: \"kubernetes.io/projected/01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd-kube-api-access-4pztz\") pod \"nova-cell1-db-create-69jjr\" (UID: \"01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd\") " pod="openstack/nova-cell1-db-create-69jjr" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.338196 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pztz\" (UniqueName: \"kubernetes.io/projected/01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd-kube-api-access-4pztz\") pod \"nova-cell1-db-create-69jjr\" (UID: \"01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd\") " pod="openstack/nova-cell1-db-create-69jjr" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.374334 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pztz\" (UniqueName: \"kubernetes.io/projected/01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd-kube-api-access-4pztz\") pod \"nova-cell1-db-create-69jjr\" (UID: \"01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd\") " pod="openstack/nova-cell1-db-create-69jjr" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.435443 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-69jjr" Oct 08 14:23:05 crc kubenswrapper[4789]: I1008 14:23:05.972113 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xll5s"] Oct 08 14:23:06 crc kubenswrapper[4789]: I1008 14:23:06.075817 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xll5s" event={"ID":"150d7f22-fe1e-47cf-bf5e-7f97473446b8","Type":"ContainerStarted","Data":"e9c91dd374766c21a61db5c6885e61c72e463b9782ec26c42da699339d8d9090"} Oct 08 14:23:06 crc kubenswrapper[4789]: I1008 14:23:06.144647 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-2wfw9"] Oct 08 14:23:06 crc kubenswrapper[4789]: I1008 14:23:06.303936 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-69jjr"] Oct 08 14:23:06 crc kubenswrapper[4789]: I1008 14:23:06.454330 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 14:23:06 crc kubenswrapper[4789]: I1008 14:23:06.536891 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 14:23:06 crc kubenswrapper[4789]: I1008 14:23:06.584149 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:23:06 crc kubenswrapper[4789]: I1008 14:23:06.655878 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d44448797-vjpbw"] Oct 08 14:23:06 crc kubenswrapper[4789]: I1008 14:23:06.668861 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" podUID="c16c0be1-2f8d-4d4e-b045-1a31a3fca002" containerName="dnsmasq-dns" containerID="cri-o://e33faf1008389661c6012ab1c47b59c3d7be4fc154f831671553bee85ceb1e59" gracePeriod=10 Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.141475 4789 generic.go:334] "Generic (PLEG): container finished" podID="c16c0be1-2f8d-4d4e-b045-1a31a3fca002" containerID="e33faf1008389661c6012ab1c47b59c3d7be4fc154f831671553bee85ceb1e59" exitCode=0 Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.141793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" event={"ID":"c16c0be1-2f8d-4d4e-b045-1a31a3fca002","Type":"ContainerDied","Data":"e33faf1008389661c6012ab1c47b59c3d7be4fc154f831671553bee85ceb1e59"} Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.148835 4789 generic.go:334] "Generic (PLEG): container finished" podID="150d7f22-fe1e-47cf-bf5e-7f97473446b8" containerID="4855c397e082720541e29543832b80834d6c77a86884b6f30d2c64d033f7c2de" exitCode=0 Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.148909 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xll5s" event={"ID":"150d7f22-fe1e-47cf-bf5e-7f97473446b8","Type":"ContainerDied","Data":"4855c397e082720541e29543832b80834d6c77a86884b6f30d2c64d033f7c2de"} Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.172488 4789 generic.go:334] "Generic (PLEG): container finished" podID="01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd" containerID="c0252b016c5a8f9976248ef4ed38cc13853d5b34265149fb7797f1bf250e2874" exitCode=0 Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.172585 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-69jjr" event={"ID":"01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd","Type":"ContainerDied","Data":"c0252b016c5a8f9976248ef4ed38cc13853d5b34265149fb7797f1bf250e2874"} Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.172621 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-69jjr" event={"ID":"01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd","Type":"ContainerStarted","Data":"88e8dcabce431c14a0ae07d9fab5d0074c5475378149d783bfe4d21402a306c4"} Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.188179 4789 generic.go:334] "Generic (PLEG): container finished" podID="05a2fc22-aea2-42aa-810a-98bbf5ba7c98" containerID="88d3dc78c1125ad9e5c403a1c650c6d4011e790180ddbfb07cd5d56050d2a189" exitCode=0 Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.188614 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c611766d-13b9-4814-bff6-0932c986de51" containerName="cinder-scheduler" containerID="cri-o://d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba" gracePeriod=30 Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.189378 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2wfw9" event={"ID":"05a2fc22-aea2-42aa-810a-98bbf5ba7c98","Type":"ContainerDied","Data":"88d3dc78c1125ad9e5c403a1c650c6d4011e790180ddbfb07cd5d56050d2a189"} Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.189479 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2wfw9" event={"ID":"05a2fc22-aea2-42aa-810a-98bbf5ba7c98","Type":"ContainerStarted","Data":"6ae3095c9c7e064d72de0f23d595a61bd58315acbb8f705787e4d9f9000db9cd"} Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.189815 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c611766d-13b9-4814-bff6-0932c986de51" containerName="probe" containerID="cri-o://142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179" gracePeriod=30 Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.446493 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.548973 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-swift-storage-0\") pod \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.549166 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-svc\") pod \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.549192 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-nb\") pod \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.549248 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-sb\") pod \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.549307 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fnlx\" (UniqueName: \"kubernetes.io/projected/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-kube-api-access-8fnlx\") pod \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.549361 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-config\") pod \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\" (UID: \"c16c0be1-2f8d-4d4e-b045-1a31a3fca002\") " Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.565802 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-kube-api-access-8fnlx" (OuterVolumeSpecName: "kube-api-access-8fnlx") pod "c16c0be1-2f8d-4d4e-b045-1a31a3fca002" (UID: "c16c0be1-2f8d-4d4e-b045-1a31a3fca002"). InnerVolumeSpecName "kube-api-access-8fnlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.608871 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c16c0be1-2f8d-4d4e-b045-1a31a3fca002" (UID: "c16c0be1-2f8d-4d4e-b045-1a31a3fca002"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.627943 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c16c0be1-2f8d-4d4e-b045-1a31a3fca002" (UID: "c16c0be1-2f8d-4d4e-b045-1a31a3fca002"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.651297 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.651331 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.651341 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fnlx\" (UniqueName: \"kubernetes.io/projected/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-kube-api-access-8fnlx\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.657431 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-config" (OuterVolumeSpecName: "config") pod "c16c0be1-2f8d-4d4e-b045-1a31a3fca002" (UID: "c16c0be1-2f8d-4d4e-b045-1a31a3fca002"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.732344 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c16c0be1-2f8d-4d4e-b045-1a31a3fca002" (UID: "c16c0be1-2f8d-4d4e-b045-1a31a3fca002"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.734477 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c16c0be1-2f8d-4d4e-b045-1a31a3fca002" (UID: "c16c0be1-2f8d-4d4e-b045-1a31a3fca002"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.756016 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.756047 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:07 crc kubenswrapper[4789]: I1008 14:23:07.756058 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c16c0be1-2f8d-4d4e-b045-1a31a3fca002-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.204740 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" event={"ID":"c16c0be1-2f8d-4d4e-b045-1a31a3fca002","Type":"ContainerDied","Data":"e3bccd422af36397e5cfd0630258bf12891a94befcd3b24b2b3cf8c5c2b699a3"} Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.205112 4789 scope.go:117] "RemoveContainer" containerID="e33faf1008389661c6012ab1c47b59c3d7be4fc154f831671553bee85ceb1e59" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.205256 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d44448797-vjpbw" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.214019 4789 generic.go:334] "Generic (PLEG): container finished" podID="c611766d-13b9-4814-bff6-0932c986de51" containerID="142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179" exitCode=0 Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.214194 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c611766d-13b9-4814-bff6-0932c986de51","Type":"ContainerDied","Data":"142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179"} Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.263967 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d44448797-vjpbw"] Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.273798 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d44448797-vjpbw"] Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.303185 4789 scope.go:117] "RemoveContainer" containerID="a22f20e7f6e8dffda6ffeed1e9d6389536e12cccd3e2e29c5f7451e23e77f5f4" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.748349 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xll5s" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.750799 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c16c0be1-2f8d-4d4e-b045-1a31a3fca002" path="/var/lib/kubelet/pods/c16c0be1-2f8d-4d4e-b045-1a31a3fca002/volumes" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.879159 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq88t\" (UniqueName: \"kubernetes.io/projected/150d7f22-fe1e-47cf-bf5e-7f97473446b8-kube-api-access-lq88t\") pod \"150d7f22-fe1e-47cf-bf5e-7f97473446b8\" (UID: \"150d7f22-fe1e-47cf-bf5e-7f97473446b8\") " Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.892879 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/150d7f22-fe1e-47cf-bf5e-7f97473446b8-kube-api-access-lq88t" (OuterVolumeSpecName: "kube-api-access-lq88t") pod "150d7f22-fe1e-47cf-bf5e-7f97473446b8" (UID: "150d7f22-fe1e-47cf-bf5e-7f97473446b8"). InnerVolumeSpecName "kube-api-access-lq88t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.936477 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.956154 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-59fddc7b48-jvk6w" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.981253 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq88t\" (UniqueName: \"kubernetes.io/projected/150d7f22-fe1e-47cf-bf5e-7f97473446b8-kube-api-access-lq88t\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:08 crc kubenswrapper[4789]: I1008 14:23:08.998675 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-69jjr" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.013905 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-9cc5c8f44-4k7vr"] Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.014383 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-9cc5c8f44-4k7vr" podUID="094807cf-ec15-4149-9967-5c52e35f93c2" containerName="barbican-api-log" containerID="cri-o://5197cad88885c384fd6f2daddd3438d9a12fd275145e51ed550c04fe297700ef" gracePeriod=30 Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.015358 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-9cc5c8f44-4k7vr" podUID="094807cf-ec15-4149-9967-5c52e35f93c2" containerName="barbican-api" containerID="cri-o://e98ee6b5a44a0d4b42e2e77219fdfff8c7fd27d9c3e3e41e6edb6671fba0632e" gracePeriod=30 Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.028906 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2wfw9" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.092544 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pztz\" (UniqueName: \"kubernetes.io/projected/01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd-kube-api-access-4pztz\") pod \"01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd\" (UID: \"01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd\") " Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.102704 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd-kube-api-access-4pztz" (OuterVolumeSpecName: "kube-api-access-4pztz") pod "01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd" (UID: "01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd"). InnerVolumeSpecName "kube-api-access-4pztz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.194745 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfsw4\" (UniqueName: \"kubernetes.io/projected/05a2fc22-aea2-42aa-810a-98bbf5ba7c98-kube-api-access-kfsw4\") pod \"05a2fc22-aea2-42aa-810a-98bbf5ba7c98\" (UID: \"05a2fc22-aea2-42aa-810a-98bbf5ba7c98\") " Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.195245 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pztz\" (UniqueName: \"kubernetes.io/projected/01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd-kube-api-access-4pztz\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.202168 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05a2fc22-aea2-42aa-810a-98bbf5ba7c98-kube-api-access-kfsw4" (OuterVolumeSpecName: "kube-api-access-kfsw4") pod "05a2fc22-aea2-42aa-810a-98bbf5ba7c98" (UID: "05a2fc22-aea2-42aa-810a-98bbf5ba7c98"). InnerVolumeSpecName "kube-api-access-kfsw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.230144 4789 generic.go:334] "Generic (PLEG): container finished" podID="094807cf-ec15-4149-9967-5c52e35f93c2" containerID="5197cad88885c384fd6f2daddd3438d9a12fd275145e51ed550c04fe297700ef" exitCode=143 Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.230211 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9cc5c8f44-4k7vr" event={"ID":"094807cf-ec15-4149-9967-5c52e35f93c2","Type":"ContainerDied","Data":"5197cad88885c384fd6f2daddd3438d9a12fd275145e51ed550c04fe297700ef"} Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.235237 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xll5s" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.235666 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xll5s" event={"ID":"150d7f22-fe1e-47cf-bf5e-7f97473446b8","Type":"ContainerDied","Data":"e9c91dd374766c21a61db5c6885e61c72e463b9782ec26c42da699339d8d9090"} Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.235701 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9c91dd374766c21a61db5c6885e61c72e463b9782ec26c42da699339d8d9090" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.261404 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-69jjr" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.262162 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-69jjr" event={"ID":"01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd","Type":"ContainerDied","Data":"88e8dcabce431c14a0ae07d9fab5d0074c5475378149d783bfe4d21402a306c4"} Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.262219 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88e8dcabce431c14a0ae07d9fab5d0074c5475378149d783bfe4d21402a306c4" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.269265 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2wfw9" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.270912 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2wfw9" event={"ID":"05a2fc22-aea2-42aa-810a-98bbf5ba7c98","Type":"ContainerDied","Data":"6ae3095c9c7e064d72de0f23d595a61bd58315acbb8f705787e4d9f9000db9cd"} Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.270948 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ae3095c9c7e064d72de0f23d595a61bd58315acbb8f705787e4d9f9000db9cd" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.296656 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfsw4\" (UniqueName: \"kubernetes.io/projected/05a2fc22-aea2-42aa-810a-98bbf5ba7c98-kube-api-access-kfsw4\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.602169 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.704785 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data-custom\") pod \"c611766d-13b9-4814-bff6-0932c986de51\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.704879 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67cd2\" (UniqueName: \"kubernetes.io/projected/c611766d-13b9-4814-bff6-0932c986de51-kube-api-access-67cd2\") pod \"c611766d-13b9-4814-bff6-0932c986de51\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.704904 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data\") pod \"c611766d-13b9-4814-bff6-0932c986de51\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.704919 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-scripts\") pod \"c611766d-13b9-4814-bff6-0932c986de51\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.704959 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-combined-ca-bundle\") pod \"c611766d-13b9-4814-bff6-0932c986de51\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.705053 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c611766d-13b9-4814-bff6-0932c986de51-etc-machine-id\") pod \"c611766d-13b9-4814-bff6-0932c986de51\" (UID: \"c611766d-13b9-4814-bff6-0932c986de51\") " Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.705314 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c611766d-13b9-4814-bff6-0932c986de51-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c611766d-13b9-4814-bff6-0932c986de51" (UID: "c611766d-13b9-4814-bff6-0932c986de51"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.720638 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c611766d-13b9-4814-bff6-0932c986de51-kube-api-access-67cd2" (OuterVolumeSpecName: "kube-api-access-67cd2") pod "c611766d-13b9-4814-bff6-0932c986de51" (UID: "c611766d-13b9-4814-bff6-0932c986de51"). InnerVolumeSpecName "kube-api-access-67cd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.728236 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c611766d-13b9-4814-bff6-0932c986de51" (UID: "c611766d-13b9-4814-bff6-0932c986de51"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.730393 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-scripts" (OuterVolumeSpecName: "scripts") pod "c611766d-13b9-4814-bff6-0932c986de51" (UID: "c611766d-13b9-4814-bff6-0932c986de51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.797067 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.807229 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67cd2\" (UniqueName: \"kubernetes.io/projected/c611766d-13b9-4814-bff6-0932c986de51-kube-api-access-67cd2\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.807285 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.807299 4789 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c611766d-13b9-4814-bff6-0932c986de51-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.807313 4789 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.822676 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c611766d-13b9-4814-bff6-0932c986de51" (UID: "c611766d-13b9-4814-bff6-0932c986de51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.840838 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.907267 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data" (OuterVolumeSpecName: "config-data") pod "c611766d-13b9-4814-bff6-0932c986de51" (UID: "c611766d-13b9-4814-bff6-0932c986de51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.911139 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:09 crc kubenswrapper[4789]: I1008 14:23:09.911181 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c611766d-13b9-4814-bff6-0932c986de51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.280354 4789 generic.go:334] "Generic (PLEG): container finished" podID="c611766d-13b9-4814-bff6-0932c986de51" containerID="d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba" exitCode=0 Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.281272 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c611766d-13b9-4814-bff6-0932c986de51","Type":"ContainerDied","Data":"d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba"} Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.281299 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c611766d-13b9-4814-bff6-0932c986de51","Type":"ContainerDied","Data":"77952a8d887f5c6fe9956b93f334f2f9b9bcf140a6d2521e1573f72b07e5f319"} Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.281314 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.281333 4789 scope.go:117] "RemoveContainer" containerID="142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.281553 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.322385 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.327764 4789 scope.go:117] "RemoveContainer" containerID="d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.331037 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.356528 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367046 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 14:23:10 crc kubenswrapper[4789]: E1008 14:23:10.367513 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c611766d-13b9-4814-bff6-0932c986de51" containerName="probe" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367532 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c611766d-13b9-4814-bff6-0932c986de51" containerName="probe" Oct 08 14:23:10 crc kubenswrapper[4789]: E1008 14:23:10.367554 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16c0be1-2f8d-4d4e-b045-1a31a3fca002" containerName="dnsmasq-dns" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367561 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16c0be1-2f8d-4d4e-b045-1a31a3fca002" containerName="dnsmasq-dns" Oct 08 14:23:10 crc kubenswrapper[4789]: E1008 14:23:10.367576 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05a2fc22-aea2-42aa-810a-98bbf5ba7c98" containerName="mariadb-database-create" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367584 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="05a2fc22-aea2-42aa-810a-98bbf5ba7c98" containerName="mariadb-database-create" Oct 08 14:23:10 crc kubenswrapper[4789]: E1008 14:23:10.367611 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c611766d-13b9-4814-bff6-0932c986de51" containerName="cinder-scheduler" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367617 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c611766d-13b9-4814-bff6-0932c986de51" containerName="cinder-scheduler" Oct 08 14:23:10 crc kubenswrapper[4789]: E1008 14:23:10.367634 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150d7f22-fe1e-47cf-bf5e-7f97473446b8" containerName="mariadb-database-create" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367641 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="150d7f22-fe1e-47cf-bf5e-7f97473446b8" containerName="mariadb-database-create" Oct 08 14:23:10 crc kubenswrapper[4789]: E1008 14:23:10.367656 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd" containerName="mariadb-database-create" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367663 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd" containerName="mariadb-database-create" Oct 08 14:23:10 crc kubenswrapper[4789]: E1008 14:23:10.367677 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16c0be1-2f8d-4d4e-b045-1a31a3fca002" containerName="init" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367683 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16c0be1-2f8d-4d4e-b045-1a31a3fca002" containerName="init" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367960 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16c0be1-2f8d-4d4e-b045-1a31a3fca002" containerName="dnsmasq-dns" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367971 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="05a2fc22-aea2-42aa-810a-98bbf5ba7c98" containerName="mariadb-database-create" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.367979 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c611766d-13b9-4814-bff6-0932c986de51" containerName="probe" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.368007 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd" containerName="mariadb-database-create" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.368026 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c611766d-13b9-4814-bff6-0932c986de51" containerName="cinder-scheduler" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.368042 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="150d7f22-fe1e-47cf-bf5e-7f97473446b8" containerName="mariadb-database-create" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.372331 4789 scope.go:117] "RemoveContainer" containerID="142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179" Oct 08 14:23:10 crc kubenswrapper[4789]: E1008 14:23:10.373267 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179\": container with ID starting with 142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179 not found: ID does not exist" containerID="142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.373303 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179"} err="failed to get container status \"142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179\": rpc error: code = NotFound desc = could not find container \"142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179\": container with ID starting with 142980bd0b398e16a338ebd3c22689a2a0e194c6c45190dc3f298afdb9664179 not found: ID does not exist" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.373327 4789 scope.go:117] "RemoveContainer" containerID="d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba" Oct 08 14:23:10 crc kubenswrapper[4789]: E1008 14:23:10.380165 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba\": container with ID starting with d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba not found: ID does not exist" containerID="d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.380237 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba"} err="failed to get container status \"d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba\": rpc error: code = NotFound desc = could not find container \"d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba\": container with ID starting with d5c1649ca43c02926372bd0881515b1e38ef4edc968d1f65a06078d223acf5ba not found: ID does not exist" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.382392 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.382555 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.386470 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.522344 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4717d784-db56-4840-8b93-9be19efb814b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.522647 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-scripts\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.522724 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.522825 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-config-data\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.522921 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.523059 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szmss\" (UniqueName: \"kubernetes.io/projected/4717d784-db56-4840-8b93-9be19efb814b-kube-api-access-szmss\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.624792 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szmss\" (UniqueName: \"kubernetes.io/projected/4717d784-db56-4840-8b93-9be19efb814b-kube-api-access-szmss\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.624858 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4717d784-db56-4840-8b93-9be19efb814b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.624919 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-scripts\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.624946 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.625004 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-config-data\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.625014 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4717d784-db56-4840-8b93-9be19efb814b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.625048 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.629484 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-scripts\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.633050 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.633089 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-config-data\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.633498 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4717d784-db56-4840-8b93-9be19efb814b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.643092 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szmss\" (UniqueName: \"kubernetes.io/projected/4717d784-db56-4840-8b93-9be19efb814b-kube-api-access-szmss\") pod \"cinder-scheduler-0\" (UID: \"4717d784-db56-4840-8b93-9be19efb814b\") " pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.708231 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 14:23:10 crc kubenswrapper[4789]: I1008 14:23:10.748783 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c611766d-13b9-4814-bff6-0932c986de51" path="/var/lib/kubelet/pods/c611766d-13b9-4814-bff6-0932c986de51/volumes" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.287479 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.335863 4789 generic.go:334] "Generic (PLEG): container finished" podID="094807cf-ec15-4149-9967-5c52e35f93c2" containerID="e98ee6b5a44a0d4b42e2e77219fdfff8c7fd27d9c3e3e41e6edb6671fba0632e" exitCode=0 Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.335918 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9cc5c8f44-4k7vr" event={"ID":"094807cf-ec15-4149-9967-5c52e35f93c2","Type":"ContainerDied","Data":"e98ee6b5a44a0d4b42e2e77219fdfff8c7fd27d9c3e3e41e6edb6671fba0632e"} Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.351906 4789 generic.go:334] "Generic (PLEG): container finished" podID="5d62cf1d-f82b-4eed-8500-2614367a4e3d" containerID="25383f9b103897430455bdc8eb3b2c62de60a49ab00c45d36fae689fc954e7c7" exitCode=0 Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.352450 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5c9mc" event={"ID":"5d62cf1d-f82b-4eed-8500-2614367a4e3d","Type":"ContainerDied","Data":"25383f9b103897430455bdc8eb3b2c62de60a49ab00c45d36fae689fc954e7c7"} Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.545426 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.635107 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.177:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.662736 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prxrr\" (UniqueName: \"kubernetes.io/projected/094807cf-ec15-4149-9967-5c52e35f93c2-kube-api-access-prxrr\") pod \"094807cf-ec15-4149-9967-5c52e35f93c2\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.663203 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/094807cf-ec15-4149-9967-5c52e35f93c2-logs\") pod \"094807cf-ec15-4149-9967-5c52e35f93c2\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.663235 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-combined-ca-bundle\") pod \"094807cf-ec15-4149-9967-5c52e35f93c2\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.663267 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data\") pod \"094807cf-ec15-4149-9967-5c52e35f93c2\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.663304 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data-custom\") pod \"094807cf-ec15-4149-9967-5c52e35f93c2\" (UID: \"094807cf-ec15-4149-9967-5c52e35f93c2\") " Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.663600 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/094807cf-ec15-4149-9967-5c52e35f93c2-logs" (OuterVolumeSpecName: "logs") pod "094807cf-ec15-4149-9967-5c52e35f93c2" (UID: "094807cf-ec15-4149-9967-5c52e35f93c2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.663834 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/094807cf-ec15-4149-9967-5c52e35f93c2-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.667087 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "094807cf-ec15-4149-9967-5c52e35f93c2" (UID: "094807cf-ec15-4149-9967-5c52e35f93c2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.668461 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/094807cf-ec15-4149-9967-5c52e35f93c2-kube-api-access-prxrr" (OuterVolumeSpecName: "kube-api-access-prxrr") pod "094807cf-ec15-4149-9967-5c52e35f93c2" (UID: "094807cf-ec15-4149-9967-5c52e35f93c2"). InnerVolumeSpecName "kube-api-access-prxrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.695287 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "094807cf-ec15-4149-9967-5c52e35f93c2" (UID: "094807cf-ec15-4149-9967-5c52e35f93c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.727717 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data" (OuterVolumeSpecName: "config-data") pod "094807cf-ec15-4149-9967-5c52e35f93c2" (UID: "094807cf-ec15-4149-9967-5c52e35f93c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.766205 4789 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.766251 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prxrr\" (UniqueName: \"kubernetes.io/projected/094807cf-ec15-4149-9967-5c52e35f93c2-kube-api-access-prxrr\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.766265 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:11 crc kubenswrapper[4789]: I1008 14:23:11.766275 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/094807cf-ec15-4149-9967-5c52e35f93c2-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.365510 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9cc5c8f44-4k7vr" event={"ID":"094807cf-ec15-4149-9967-5c52e35f93c2","Type":"ContainerDied","Data":"e73d55e2083d59c0e25e7ce5d100e7e3c529f52c60511d8ee591f62638671ace"} Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.365851 4789 scope.go:117] "RemoveContainer" containerID="e98ee6b5a44a0d4b42e2e77219fdfff8c7fd27d9c3e3e41e6edb6671fba0632e" Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.365534 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9cc5c8f44-4k7vr" Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.368260 4789 generic.go:334] "Generic (PLEG): container finished" podID="0b096b2c-0d8d-44bc-9275-e712b6e622b8" containerID="c1bf52812a5d2149fb993cd3676c09b67d4e4fdde81d2f4ee339b9debb591dc0" exitCode=0 Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.368335 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p54dp" event={"ID":"0b096b2c-0d8d-44bc-9275-e712b6e622b8","Type":"ContainerDied","Data":"c1bf52812a5d2149fb993cd3676c09b67d4e4fdde81d2f4ee339b9debb591dc0"} Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.370536 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4717d784-db56-4840-8b93-9be19efb814b","Type":"ContainerStarted","Data":"2b77fadc7207c55187b93b2002914aa24665fb2d5db25453b96bb4c9bd79b4fe"} Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.370580 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4717d784-db56-4840-8b93-9be19efb814b","Type":"ContainerStarted","Data":"ae4e2b202e5fae6910b034d9ec57fa644a9a023a7693ea45d3dba76a5ce0adc2"} Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.504944 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-9cc5c8f44-4k7vr"] Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.511500 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-9cc5c8f44-4k7vr"] Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.520920 4789 scope.go:117] "RemoveContainer" containerID="5197cad88885c384fd6f2daddd3438d9a12fd275145e51ed550c04fe297700ef" Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.760418 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="094807cf-ec15-4149-9967-5c52e35f93c2" path="/var/lib/kubelet/pods/094807cf-ec15-4149-9967-5c52e35f93c2/volumes" Oct 08 14:23:12 crc kubenswrapper[4789]: I1008 14:23:12.937243 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5c9mc" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.097179 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txzxz\" (UniqueName: \"kubernetes.io/projected/5d62cf1d-f82b-4eed-8500-2614367a4e3d-kube-api-access-txzxz\") pod \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.097356 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-db-sync-config-data\") pod \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.097410 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-combined-ca-bundle\") pod \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.097536 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-config-data\") pod \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\" (UID: \"5d62cf1d-f82b-4eed-8500-2614367a4e3d\") " Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.113079 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d62cf1d-f82b-4eed-8500-2614367a4e3d-kube-api-access-txzxz" (OuterVolumeSpecName: "kube-api-access-txzxz") pod "5d62cf1d-f82b-4eed-8500-2614367a4e3d" (UID: "5d62cf1d-f82b-4eed-8500-2614367a4e3d"). InnerVolumeSpecName "kube-api-access-txzxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.113789 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5d62cf1d-f82b-4eed-8500-2614367a4e3d" (UID: "5d62cf1d-f82b-4eed-8500-2614367a4e3d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.137891 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d62cf1d-f82b-4eed-8500-2614367a4e3d" (UID: "5d62cf1d-f82b-4eed-8500-2614367a4e3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.168298 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-config-data" (OuterVolumeSpecName: "config-data") pod "5d62cf1d-f82b-4eed-8500-2614367a4e3d" (UID: "5d62cf1d-f82b-4eed-8500-2614367a4e3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.199795 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.199835 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.199844 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txzxz\" (UniqueName: \"kubernetes.io/projected/5d62cf1d-f82b-4eed-8500-2614367a4e3d-kube-api-access-txzxz\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.199857 4789 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d62cf1d-f82b-4eed-8500-2614367a4e3d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.380935 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5c9mc" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.380939 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5c9mc" event={"ID":"5d62cf1d-f82b-4eed-8500-2614367a4e3d","Type":"ContainerDied","Data":"19b87444e11c1e7411fdfde5ceef1e8091450a6c9dc2973af2e2b4cb59352ca2"} Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.381030 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19b87444e11c1e7411fdfde5ceef1e8091450a6c9dc2973af2e2b4cb59352ca2" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.387333 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4717d784-db56-4840-8b93-9be19efb814b","Type":"ContainerStarted","Data":"5958c4628a6a95b086bf51c99a1cb0f666f5f28fd90a5a2160ae451719613d6f"} Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.411861 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.411821942 podStartE2EDuration="3.411821942s" podCreationTimestamp="2025-10-08 14:23:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:13.410093975 +0000 UTC m=+1333.316841467" watchObservedRunningTime="2025-10-08 14:23:13.411821942 +0000 UTC m=+1333.318569434" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.853033 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p54dp" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.887039 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f479f5f75-vtsl6"] Oct 08 14:23:13 crc kubenswrapper[4789]: E1008 14:23:13.887529 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094807cf-ec15-4149-9967-5c52e35f93c2" containerName="barbican-api" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.887566 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="094807cf-ec15-4149-9967-5c52e35f93c2" containerName="barbican-api" Oct 08 14:23:13 crc kubenswrapper[4789]: E1008 14:23:13.887591 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094807cf-ec15-4149-9967-5c52e35f93c2" containerName="barbican-api-log" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.887600 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="094807cf-ec15-4149-9967-5c52e35f93c2" containerName="barbican-api-log" Oct 08 14:23:13 crc kubenswrapper[4789]: E1008 14:23:13.887641 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b096b2c-0d8d-44bc-9275-e712b6e622b8" containerName="neutron-db-sync" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.887654 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b096b2c-0d8d-44bc-9275-e712b6e622b8" containerName="neutron-db-sync" Oct 08 14:23:13 crc kubenswrapper[4789]: E1008 14:23:13.887678 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d62cf1d-f82b-4eed-8500-2614367a4e3d" containerName="glance-db-sync" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.887688 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d62cf1d-f82b-4eed-8500-2614367a4e3d" containerName="glance-db-sync" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.887939 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b096b2c-0d8d-44bc-9275-e712b6e622b8" containerName="neutron-db-sync" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.887969 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="094807cf-ec15-4149-9967-5c52e35f93c2" containerName="barbican-api-log" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.888094 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="094807cf-ec15-4149-9967-5c52e35f93c2" containerName="barbican-api" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.888121 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d62cf1d-f82b-4eed-8500-2614367a4e3d" containerName="glance-db-sync" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.889421 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.916338 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f479f5f75-vtsl6"] Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.917630 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-489x7\" (UniqueName: \"kubernetes.io/projected/bfb63537-19e3-4a3c-a2b2-2b708c79fded-kube-api-access-489x7\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.917689 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-svc\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.917842 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-nb\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.917874 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-swift-storage-0\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.917942 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-sb\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:13 crc kubenswrapper[4789]: I1008 14:23:13.917974 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-config\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.019182 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5txnt\" (UniqueName: \"kubernetes.io/projected/0b096b2c-0d8d-44bc-9275-e712b6e622b8-kube-api-access-5txnt\") pod \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.019336 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-config\") pod \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.019403 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-combined-ca-bundle\") pod \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\" (UID: \"0b096b2c-0d8d-44bc-9275-e712b6e622b8\") " Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.019729 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-svc\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.019831 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-nb\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.019855 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-swift-storage-0\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.019919 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-sb\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.020161 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-config\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.020276 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-489x7\" (UniqueName: \"kubernetes.io/projected/bfb63537-19e3-4a3c-a2b2-2b708c79fded-kube-api-access-489x7\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.020613 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-svc\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.020706 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-nb\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.021588 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-swift-storage-0\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.023115 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-sb\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.024149 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-config\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.026864 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b096b2c-0d8d-44bc-9275-e712b6e622b8-kube-api-access-5txnt" (OuterVolumeSpecName: "kube-api-access-5txnt") pod "0b096b2c-0d8d-44bc-9275-e712b6e622b8" (UID: "0b096b2c-0d8d-44bc-9275-e712b6e622b8"). InnerVolumeSpecName "kube-api-access-5txnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.058210 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b096b2c-0d8d-44bc-9275-e712b6e622b8" (UID: "0b096b2c-0d8d-44bc-9275-e712b6e622b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.059375 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-489x7\" (UniqueName: \"kubernetes.io/projected/bfb63537-19e3-4a3c-a2b2-2b708c79fded-kube-api-access-489x7\") pod \"dnsmasq-dns-6f479f5f75-vtsl6\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.081143 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-config" (OuterVolumeSpecName: "config") pod "0b096b2c-0d8d-44bc-9275-e712b6e622b8" (UID: "0b096b2c-0d8d-44bc-9275-e712b6e622b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.122030 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5txnt\" (UniqueName: \"kubernetes.io/projected/0b096b2c-0d8d-44bc-9275-e712b6e622b8-kube-api-access-5txnt\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.122076 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.122090 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b096b2c-0d8d-44bc-9275-e712b6e622b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.256388 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.279279 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.418217 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-p54dp" event={"ID":"0b096b2c-0d8d-44bc-9275-e712b6e622b8","Type":"ContainerDied","Data":"10fda8566c16914cc6daa7620204ac1ae3b0a09b92dd436053da2afadbf4e3de"} Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.418278 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10fda8566c16914cc6daa7620204ac1ae3b0a09b92dd436053da2afadbf4e3de" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.418377 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-p54dp" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.593907 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f479f5f75-vtsl6"] Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.633125 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bb6987789-gwcfk"] Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.638408 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.672371 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bb6987789-gwcfk"] Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.770448 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.772176 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.778064 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8687c9c966-lnx5h"] Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.780428 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.782018 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rckbp" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.782303 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.782329 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.784098 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.795113 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.805544 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xnlqp" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.805602 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.805689 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.833196 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8687c9c966-lnx5h"] Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.850211 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-swift-storage-0\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.850279 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-nb\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.850462 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-sb\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.850515 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-svc\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.850550 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chfd8\" (UniqueName: \"kubernetes.io/projected/7cbf3178-8d77-47d9-beb0-804732748d97-kube-api-access-chfd8\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.850581 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-config\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.951253 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.170:3000/\": dial tcp 10.217.0.170:3000: connect: connection refused" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.971187 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm6jw\" (UniqueName: \"kubernetes.io/projected/09623c61-6d40-4dd4-8f66-e9b398e9747f-kube-api-access-xm6jw\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972300 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972389 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-combined-ca-bundle\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972459 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-scripts\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972626 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-config\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972698 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-sb\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972784 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-svc\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972817 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chfd8\" (UniqueName: \"kubernetes.io/projected/7cbf3178-8d77-47d9-beb0-804732748d97-kube-api-access-chfd8\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972859 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-ovndb-tls-certs\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972891 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-config\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972915 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.972953 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j67tm\" (UniqueName: \"kubernetes.io/projected/1e5167e7-8bb8-4580-b667-d0f5083b156b-kube-api-access-j67tm\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.973010 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-logs\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.973036 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-swift-storage-0\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.973072 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.973096 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-nb\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.973116 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-config-data\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.973178 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-httpd-config\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.974149 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-sb\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.974671 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-svc\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.975525 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-config\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.976213 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-nb\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:14 crc kubenswrapper[4789]: I1008 14:23:14.979821 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-swift-storage-0\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.068432 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chfd8\" (UniqueName: \"kubernetes.io/projected/7cbf3178-8d77-47d9-beb0-804732748d97-kube-api-access-chfd8\") pod \"dnsmasq-dns-bb6987789-gwcfk\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.097402 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-b687-account-create-4s2zx"] Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102435 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-ovndb-tls-certs\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102495 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102534 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j67tm\" (UniqueName: \"kubernetes.io/projected/1e5167e7-8bb8-4580-b667-d0f5083b156b-kube-api-access-j67tm\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102561 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-logs\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102595 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102615 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-config-data\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102658 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-httpd-config\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102717 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm6jw\" (UniqueName: \"kubernetes.io/projected/09623c61-6d40-4dd4-8f66-e9b398e9747f-kube-api-access-xm6jw\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102739 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102775 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-combined-ca-bundle\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102819 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-scripts\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.102870 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-config\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.118108 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-config-data\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.118542 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-logs\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.118862 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.119408 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.127642 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b687-account-create-4s2zx" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.127927 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-config\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.129982 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.130301 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.140939 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-scripts\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.148731 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-combined-ca-bundle\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.150769 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-httpd-config\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.151713 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-ovndb-tls-certs\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.183172 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b687-account-create-4s2zx"] Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.199224 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j67tm\" (UniqueName: \"kubernetes.io/projected/1e5167e7-8bb8-4580-b667-d0f5083b156b-kube-api-access-j67tm\") pod \"neutron-8687c9c966-lnx5h\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.210401 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm6jw\" (UniqueName: \"kubernetes.io/projected/09623c61-6d40-4dd4-8f66-e9b398e9747f-kube-api-access-xm6jw\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.214954 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.247507 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f479f5f75-vtsl6"] Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.267116 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.268867 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.275098 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.293632 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.307184 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78gxn\" (UniqueName: \"kubernetes.io/projected/36757dfa-8de4-475b-bf51-7320d55c16c7-kube-api-access-78gxn\") pod \"nova-api-b687-account-create-4s2zx\" (UID: \"36757dfa-8de4-475b-bf51-7320d55c16c7\") " pod="openstack/nova-api-b687-account-create-4s2zx" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.324681 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.326514 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-e461-account-create-sgqdc"] Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.328225 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e461-account-create-sgqdc" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.334301 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.379043 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e461-account-create-sgqdc"] Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.412233 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.412323 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78gxn\" (UniqueName: \"kubernetes.io/projected/36757dfa-8de4-475b-bf51-7320d55c16c7-kube-api-access-78gxn\") pod \"nova-api-b687-account-create-4s2zx\" (UID: \"36757dfa-8de4-475b-bf51-7320d55c16c7\") " pod="openstack/nova-api-b687-account-create-4s2zx" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.412369 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.412442 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.420226 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-logs\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.420390 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.420468 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rttsc\" (UniqueName: \"kubernetes.io/projected/1d13191a-daae-49b6-aabe-d9c0ab0018bf-kube-api-access-rttsc\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.420494 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.433263 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.436734 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78gxn\" (UniqueName: \"kubernetes.io/projected/36757dfa-8de4-475b-bf51-7320d55c16c7-kube-api-access-78gxn\") pod \"nova-api-b687-account-create-4s2zx\" (UID: \"36757dfa-8de4-475b-bf51-7320d55c16c7\") " pod="openstack/nova-api-b687-account-create-4s2zx" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.457438 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" event={"ID":"bfb63537-19e3-4a3c-a2b2-2b708c79fded","Type":"ContainerStarted","Data":"a45107e5de257a7cc3286f4e09c2c409313b713d56accdf8a0f5d7ff837a573a"} Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.461597 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.522351 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.522818 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.522915 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6zcp\" (UniqueName: \"kubernetes.io/projected/9b985f8e-1745-4c46-9f99-d1e4978413e0-kube-api-access-l6zcp\") pod \"nova-cell1-e461-account-create-sgqdc\" (UID: \"9b985f8e-1745-4c46-9f99-d1e4978413e0\") " pod="openstack/nova-cell1-e461-account-create-sgqdc" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.523036 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-logs\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.523140 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.523223 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rttsc\" (UniqueName: \"kubernetes.io/projected/1d13191a-daae-49b6-aabe-d9c0ab0018bf-kube-api-access-rttsc\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.523279 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.523382 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.524266 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.524886 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.525536 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-logs\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.538640 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.541765 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.544695 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.561774 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rttsc\" (UniqueName: \"kubernetes.io/projected/1d13191a-daae-49b6-aabe-d9c0ab0018bf-kube-api-access-rttsc\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.607913 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b687-account-create-4s2zx" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.625343 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6zcp\" (UniqueName: \"kubernetes.io/projected/9b985f8e-1745-4c46-9f99-d1e4978413e0-kube-api-access-l6zcp\") pod \"nova-cell1-e461-account-create-sgqdc\" (UID: \"9b985f8e-1745-4c46-9f99-d1e4978413e0\") " pod="openstack/nova-cell1-e461-account-create-sgqdc" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.660610 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6zcp\" (UniqueName: \"kubernetes.io/projected/9b985f8e-1745-4c46-9f99-d1e4978413e0-kube-api-access-l6zcp\") pod \"nova-cell1-e461-account-create-sgqdc\" (UID: \"9b985f8e-1745-4c46-9f99-d1e4978413e0\") " pod="openstack/nova-cell1-e461-account-create-sgqdc" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.693610 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.700702 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e461-account-create-sgqdc" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.718497 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 14:23:15 crc kubenswrapper[4789]: I1008 14:23:15.966081 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.089730 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bb6987789-gwcfk"] Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.334195 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8687c9c966-lnx5h"] Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.367333 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b687-account-create-4s2zx"] Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.484284 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8687c9c966-lnx5h" event={"ID":"1e5167e7-8bb8-4580-b667-d0f5083b156b","Type":"ContainerStarted","Data":"c9ca7edb473dd7d8142ab7390818dcd92093ad5a6c52ba2527a4b676b043f04f"} Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.498544 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" event={"ID":"7cbf3178-8d77-47d9-beb0-804732748d97","Type":"ContainerStarted","Data":"b8e92044f3c94beedb36cb345ffadc3826d651bb3ed186b182c22681d72f6cd2"} Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.498583 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" event={"ID":"7cbf3178-8d77-47d9-beb0-804732748d97","Type":"ContainerStarted","Data":"2fefad0d6e111c73573721cdbf78c238071eb4dd4b9546dcff761034bb77165e"} Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.507959 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b687-account-create-4s2zx" event={"ID":"36757dfa-8de4-475b-bf51-7320d55c16c7","Type":"ContainerStarted","Data":"dcd589b558cb6ca4e85fc38c17da945b60799fb384ac0985e91a9296ab8fb932"} Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.514889 4789 generic.go:334] "Generic (PLEG): container finished" podID="bfb63537-19e3-4a3c-a2b2-2b708c79fded" containerID="de30aa5f0935f5bc7b92ffb840cd1665867b14658e253ea724a8d559b568698c" exitCode=0 Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.514941 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" event={"ID":"bfb63537-19e3-4a3c-a2b2-2b708c79fded","Type":"ContainerDied","Data":"de30aa5f0935f5bc7b92ffb840cd1665867b14658e253ea724a8d559b568698c"} Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.598567 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e461-account-create-sgqdc"] Oct 08 14:23:16 crc kubenswrapper[4789]: I1008 14:23:16.863498 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.042800 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.437609 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.448230 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.529733 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" event={"ID":"bfb63537-19e3-4a3c-a2b2-2b708c79fded","Type":"ContainerDied","Data":"a45107e5de257a7cc3286f4e09c2c409313b713d56accdf8a0f5d7ff837a573a"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.529781 4789 scope.go:117] "RemoveContainer" containerID="de30aa5f0935f5bc7b92ffb840cd1665867b14658e253ea724a8d559b568698c" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.529897 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f479f5f75-vtsl6" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.546176 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8687c9c966-lnx5h" event={"ID":"1e5167e7-8bb8-4580-b667-d0f5083b156b","Type":"ContainerStarted","Data":"462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.546217 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8687c9c966-lnx5h" event={"ID":"1e5167e7-8bb8-4580-b667-d0f5083b156b","Type":"ContainerStarted","Data":"1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.547374 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.555828 4789 generic.go:334] "Generic (PLEG): container finished" podID="9b985f8e-1745-4c46-9f99-d1e4978413e0" containerID="96a413984afa6267b4352c3740ff023d52b3ca71b3f4a31c97e9b86f7c99941c" exitCode=0 Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.555902 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e461-account-create-sgqdc" event={"ID":"9b985f8e-1745-4c46-9f99-d1e4978413e0","Type":"ContainerDied","Data":"96a413984afa6267b4352c3740ff023d52b3ca71b3f4a31c97e9b86f7c99941c"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.555928 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e461-account-create-sgqdc" event={"ID":"9b985f8e-1745-4c46-9f99-d1e4978413e0","Type":"ContainerStarted","Data":"785de579cbaa9300d606273f271c85c61f869181897b75ae36ccff8a3a531afe"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.559203 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d13191a-daae-49b6-aabe-d9c0ab0018bf","Type":"ContainerStarted","Data":"856e84cb67bb206160310cf995860cda03b7f8f89d964f69d0bb8569582490ce"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.562425 4789 generic.go:334] "Generic (PLEG): container finished" podID="7cbf3178-8d77-47d9-beb0-804732748d97" containerID="b8e92044f3c94beedb36cb345ffadc3826d651bb3ed186b182c22681d72f6cd2" exitCode=0 Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.562497 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" event={"ID":"7cbf3178-8d77-47d9-beb0-804732748d97","Type":"ContainerDied","Data":"b8e92044f3c94beedb36cb345ffadc3826d651bb3ed186b182c22681d72f6cd2"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.562534 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" event={"ID":"7cbf3178-8d77-47d9-beb0-804732748d97","Type":"ContainerStarted","Data":"9e226dec43528f969019b70e47ed42a9dfda14f686a49c3629e8c726c49ae94f"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.563205 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.567028 4789 generic.go:334] "Generic (PLEG): container finished" podID="36757dfa-8de4-475b-bf51-7320d55c16c7" containerID="edd9d78df92c7ffe45486e262e4ca4142bcaa9ee65566bc9f75c179f44c9d174" exitCode=0 Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.567083 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b687-account-create-4s2zx" event={"ID":"36757dfa-8de4-475b-bf51-7320d55c16c7","Type":"ContainerDied","Data":"edd9d78df92c7ffe45486e262e4ca4142bcaa9ee65566bc9f75c179f44c9d174"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.568747 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"09623c61-6d40-4dd4-8f66-e9b398e9747f","Type":"ContainerStarted","Data":"09dc8a0cf06f3f2d85816d0ab99fe3fda0c1d776690379972cd6f17ea7b577a4"} Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.586232 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-config\") pod \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.586373 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-svc\") pod \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.586533 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-nb\") pod \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.586642 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-489x7\" (UniqueName: \"kubernetes.io/projected/bfb63537-19e3-4a3c-a2b2-2b708c79fded-kube-api-access-489x7\") pod \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.586690 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-swift-storage-0\") pod \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.586727 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-sb\") pod \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\" (UID: \"bfb63537-19e3-4a3c-a2b2-2b708c79fded\") " Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.595229 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8687c9c966-lnx5h" podStartSLOduration=3.59520862 podStartE2EDuration="3.59520862s" podCreationTimestamp="2025-10-08 14:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:17.572626258 +0000 UTC m=+1337.479373760" watchObservedRunningTime="2025-10-08 14:23:17.59520862 +0000 UTC m=+1337.501956112" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.621368 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb63537-19e3-4a3c-a2b2-2b708c79fded-kube-api-access-489x7" (OuterVolumeSpecName: "kube-api-access-489x7") pod "bfb63537-19e3-4a3c-a2b2-2b708c79fded" (UID: "bfb63537-19e3-4a3c-a2b2-2b708c79fded"). InnerVolumeSpecName "kube-api-access-489x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.656966 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bfb63537-19e3-4a3c-a2b2-2b708c79fded" (UID: "bfb63537-19e3-4a3c-a2b2-2b708c79fded"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.667805 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" podStartSLOduration=3.667784226 podStartE2EDuration="3.667784226s" podCreationTimestamp="2025-10-08 14:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:17.639897811 +0000 UTC m=+1337.546645303" watchObservedRunningTime="2025-10-08 14:23:17.667784226 +0000 UTC m=+1337.574531718" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.691260 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.691298 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-489x7\" (UniqueName: \"kubernetes.io/projected/bfb63537-19e3-4a3c-a2b2-2b708c79fded-kube-api-access-489x7\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.697974 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bfb63537-19e3-4a3c-a2b2-2b708c79fded" (UID: "bfb63537-19e3-4a3c-a2b2-2b708c79fded"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.722037 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-config" (OuterVolumeSpecName: "config") pod "bfb63537-19e3-4a3c-a2b2-2b708c79fded" (UID: "bfb63537-19e3-4a3c-a2b2-2b708c79fded"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.741602 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bfb63537-19e3-4a3c-a2b2-2b708c79fded" (UID: "bfb63537-19e3-4a3c-a2b2-2b708c79fded"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.794667 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.794696 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.794704 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.801879 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bfb63537-19e3-4a3c-a2b2-2b708c79fded" (UID: "bfb63537-19e3-4a3c-a2b2-2b708c79fded"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.901073 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bfb63537-19e3-4a3c-a2b2-2b708c79fded-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.941591 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f479f5f75-vtsl6"] Oct 08 14:23:17 crc kubenswrapper[4789]: I1008 14:23:17.972914 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f479f5f75-vtsl6"] Oct 08 14:23:18 crc kubenswrapper[4789]: I1008 14:23:18.585472 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d13191a-daae-49b6-aabe-d9c0ab0018bf","Type":"ContainerStarted","Data":"effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec"} Oct 08 14:23:18 crc kubenswrapper[4789]: I1008 14:23:18.590548 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"09623c61-6d40-4dd4-8f66-e9b398e9747f","Type":"ContainerStarted","Data":"b3542835961fe21f75337e37f743e8b812c03c6d3ea91479e29f928f4774b96f"} Oct 08 14:23:18 crc kubenswrapper[4789]: I1008 14:23:18.748920 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfb63537-19e3-4a3c-a2b2-2b708c79fded" path="/var/lib/kubelet/pods/bfb63537-19e3-4a3c-a2b2-2b708c79fded/volumes" Oct 08 14:23:18 crc kubenswrapper[4789]: I1008 14:23:18.997126 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.212797 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e461-account-create-sgqdc" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.224637 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b687-account-create-4s2zx" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.330947 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78gxn\" (UniqueName: \"kubernetes.io/projected/36757dfa-8de4-475b-bf51-7320d55c16c7-kube-api-access-78gxn\") pod \"36757dfa-8de4-475b-bf51-7320d55c16c7\" (UID: \"36757dfa-8de4-475b-bf51-7320d55c16c7\") " Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.331153 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6zcp\" (UniqueName: \"kubernetes.io/projected/9b985f8e-1745-4c46-9f99-d1e4978413e0-kube-api-access-l6zcp\") pod \"9b985f8e-1745-4c46-9f99-d1e4978413e0\" (UID: \"9b985f8e-1745-4c46-9f99-d1e4978413e0\") " Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.336938 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36757dfa-8de4-475b-bf51-7320d55c16c7-kube-api-access-78gxn" (OuterVolumeSpecName: "kube-api-access-78gxn") pod "36757dfa-8de4-475b-bf51-7320d55c16c7" (UID: "36757dfa-8de4-475b-bf51-7320d55c16c7"). InnerVolumeSpecName "kube-api-access-78gxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.337102 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b985f8e-1745-4c46-9f99-d1e4978413e0-kube-api-access-l6zcp" (OuterVolumeSpecName: "kube-api-access-l6zcp") pod "9b985f8e-1745-4c46-9f99-d1e4978413e0" (UID: "9b985f8e-1745-4c46-9f99-d1e4978413e0"). InnerVolumeSpecName "kube-api-access-l6zcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.433925 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6zcp\" (UniqueName: \"kubernetes.io/projected/9b985f8e-1745-4c46-9f99-d1e4978413e0-kube-api-access-l6zcp\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.433967 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78gxn\" (UniqueName: \"kubernetes.io/projected/36757dfa-8de4-475b-bf51-7320d55c16c7-kube-api-access-78gxn\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.608679 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e461-account-create-sgqdc" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.608694 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e461-account-create-sgqdc" event={"ID":"9b985f8e-1745-4c46-9f99-d1e4978413e0","Type":"ContainerDied","Data":"785de579cbaa9300d606273f271c85c61f869181897b75ae36ccff8a3a531afe"} Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.609293 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="785de579cbaa9300d606273f271c85c61f869181897b75ae36ccff8a3a531afe" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.615250 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d13191a-daae-49b6-aabe-d9c0ab0018bf","Type":"ContainerStarted","Data":"ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251"} Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.615318 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerName="glance-log" containerID="cri-o://effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec" gracePeriod=30 Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.615377 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerName="glance-httpd" containerID="cri-o://ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251" gracePeriod=30 Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.620490 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b687-account-create-4s2zx" event={"ID":"36757dfa-8de4-475b-bf51-7320d55c16c7","Type":"ContainerDied","Data":"dcd589b558cb6ca4e85fc38c17da945b60799fb384ac0985e91a9296ab8fb932"} Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.620524 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcd589b558cb6ca4e85fc38c17da945b60799fb384ac0985e91a9296ab8fb932" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.620565 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b687-account-create-4s2zx" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.623803 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerName="glance-log" containerID="cri-o://b3542835961fe21f75337e37f743e8b812c03c6d3ea91479e29f928f4774b96f" gracePeriod=30 Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.623874 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"09623c61-6d40-4dd4-8f66-e9b398e9747f","Type":"ContainerStarted","Data":"4f2de78d9e2f4d0f6eafd577138bee33aa364de14faf849c43cca8cd6eb3e1d9"} Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.623920 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerName="glance-httpd" containerID="cri-o://4f2de78d9e2f4d0f6eafd577138bee33aa364de14faf849c43cca8cd6eb3e1d9" gracePeriod=30 Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.642749 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.642732681 podStartE2EDuration="5.642732681s" podCreationTimestamp="2025-10-08 14:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:19.639451892 +0000 UTC m=+1339.546199394" watchObservedRunningTime="2025-10-08 14:23:19.642732681 +0000 UTC m=+1339.549480173" Oct 08 14:23:19 crc kubenswrapper[4789]: I1008 14:23:19.780386 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.7803662110000005 podStartE2EDuration="6.780366211s" podCreationTimestamp="2025-10-08 14:23:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:19.695036649 +0000 UTC m=+1339.601784141" watchObservedRunningTime="2025-10-08 14:23:19.780366211 +0000 UTC m=+1339.687113703" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.596620 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.639861 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerID="ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251" exitCode=143 Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.639896 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerID="effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec" exitCode=143 Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.639965 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d13191a-daae-49b6-aabe-d9c0ab0018bf","Type":"ContainerDied","Data":"ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251"} Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.640023 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d13191a-daae-49b6-aabe-d9c0ab0018bf","Type":"ContainerDied","Data":"effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec"} Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.640035 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d13191a-daae-49b6-aabe-d9c0ab0018bf","Type":"ContainerDied","Data":"856e84cb67bb206160310cf995860cda03b7f8f89d964f69d0bb8569582490ce"} Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.640055 4789 scope.go:117] "RemoveContainer" containerID="ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.640233 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.652889 4789 generic.go:334] "Generic (PLEG): container finished" podID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerID="4f2de78d9e2f4d0f6eafd577138bee33aa364de14faf849c43cca8cd6eb3e1d9" exitCode=143 Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.652931 4789 generic.go:334] "Generic (PLEG): container finished" podID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerID="b3542835961fe21f75337e37f743e8b812c03c6d3ea91479e29f928f4774b96f" exitCode=143 Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.652959 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"09623c61-6d40-4dd4-8f66-e9b398e9747f","Type":"ContainerDied","Data":"4f2de78d9e2f4d0f6eafd577138bee33aa364de14faf849c43cca8cd6eb3e1d9"} Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.653004 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"09623c61-6d40-4dd4-8f66-e9b398e9747f","Type":"ContainerDied","Data":"b3542835961fe21f75337e37f743e8b812c03c6d3ea91479e29f928f4774b96f"} Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.672889 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-combined-ca-bundle\") pod \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.672952 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-httpd-run\") pod \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.673112 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-config-data\") pod \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.673160 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-logs\") pod \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.673221 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-scripts\") pod \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.673239 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rttsc\" (UniqueName: \"kubernetes.io/projected/1d13191a-daae-49b6-aabe-d9c0ab0018bf-kube-api-access-rttsc\") pod \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.673277 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\" (UID: \"1d13191a-daae-49b6-aabe-d9c0ab0018bf\") " Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.675620 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-logs" (OuterVolumeSpecName: "logs") pod "1d13191a-daae-49b6-aabe-d9c0ab0018bf" (UID: "1d13191a-daae-49b6-aabe-d9c0ab0018bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.679240 4789 scope.go:117] "RemoveContainer" containerID="effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.679543 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1d13191a-daae-49b6-aabe-d9c0ab0018bf" (UID: "1d13191a-daae-49b6-aabe-d9c0ab0018bf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.686925 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "1d13191a-daae-49b6-aabe-d9c0ab0018bf" (UID: "1d13191a-daae-49b6-aabe-d9c0ab0018bf"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.687128 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-scripts" (OuterVolumeSpecName: "scripts") pod "1d13191a-daae-49b6-aabe-d9c0ab0018bf" (UID: "1d13191a-daae-49b6-aabe-d9c0ab0018bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.709306 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d13191a-daae-49b6-aabe-d9c0ab0018bf-kube-api-access-rttsc" (OuterVolumeSpecName: "kube-api-access-rttsc") pod "1d13191a-daae-49b6-aabe-d9c0ab0018bf" (UID: "1d13191a-daae-49b6-aabe-d9c0ab0018bf"). InnerVolumeSpecName "kube-api-access-rttsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.722217 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d13191a-daae-49b6-aabe-d9c0ab0018bf" (UID: "1d13191a-daae-49b6-aabe-d9c0ab0018bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.775352 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.775381 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.775390 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rttsc\" (UniqueName: \"kubernetes.io/projected/1d13191a-daae-49b6-aabe-d9c0ab0018bf-kube-api-access-rttsc\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.775414 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.775424 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.775432 4789 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d13191a-daae-49b6-aabe-d9c0ab0018bf-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.801743 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.821375 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-config-data" (OuterVolumeSpecName: "config-data") pod "1d13191a-daae-49b6-aabe-d9c0ab0018bf" (UID: "1d13191a-daae-49b6-aabe-d9c0ab0018bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.826331 4789 scope.go:117] "RemoveContainer" containerID="ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251" Oct 08 14:23:20 crc kubenswrapper[4789]: E1008 14:23:20.826842 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251\": container with ID starting with ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251 not found: ID does not exist" containerID="ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.826872 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251"} err="failed to get container status \"ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251\": rpc error: code = NotFound desc = could not find container \"ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251\": container with ID starting with ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251 not found: ID does not exist" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.826899 4789 scope.go:117] "RemoveContainer" containerID="effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec" Oct 08 14:23:20 crc kubenswrapper[4789]: E1008 14:23:20.827154 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec\": container with ID starting with effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec not found: ID does not exist" containerID="effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.827176 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec"} err="failed to get container status \"effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec\": rpc error: code = NotFound desc = could not find container \"effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec\": container with ID starting with effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec not found: ID does not exist" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.827192 4789 scope.go:117] "RemoveContainer" containerID="ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.827405 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251"} err="failed to get container status \"ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251\": rpc error: code = NotFound desc = could not find container \"ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251\": container with ID starting with ab861ff8e76a5d5baa990219fe4b41bce79e04485c87fa705d26691de6deb251 not found: ID does not exist" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.827420 4789 scope.go:117] "RemoveContainer" containerID="effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.827582 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec"} err="failed to get container status \"effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec\": rpc error: code = NotFound desc = could not find container \"effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec\": container with ID starting with effa3d6eeb5f3440f478270f56c67c36666f704f4e3ef9e67f15181a601cefec not found: ID does not exist" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.877298 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13191a-daae-49b6-aabe-d9c0ab0018bf-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.877330 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:20 crc kubenswrapper[4789]: I1008 14:23:20.997414 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.006390 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.014354 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016175 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 14:23:21 crc kubenswrapper[4789]: E1008 14:23:21.016532 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerName="glance-httpd" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016546 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerName="glance-httpd" Oct 08 14:23:21 crc kubenswrapper[4789]: E1008 14:23:21.016562 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b985f8e-1745-4c46-9f99-d1e4978413e0" containerName="mariadb-account-create" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016567 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b985f8e-1745-4c46-9f99-d1e4978413e0" containerName="mariadb-account-create" Oct 08 14:23:21 crc kubenswrapper[4789]: E1008 14:23:21.016584 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36757dfa-8de4-475b-bf51-7320d55c16c7" containerName="mariadb-account-create" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016591 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="36757dfa-8de4-475b-bf51-7320d55c16c7" containerName="mariadb-account-create" Oct 08 14:23:21 crc kubenswrapper[4789]: E1008 14:23:21.016605 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerName="glance-log" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016610 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerName="glance-log" Oct 08 14:23:21 crc kubenswrapper[4789]: E1008 14:23:21.016627 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb63537-19e3-4a3c-a2b2-2b708c79fded" containerName="init" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016632 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb63537-19e3-4a3c-a2b2-2b708c79fded" containerName="init" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016814 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerName="glance-log" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016833 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b985f8e-1745-4c46-9f99-d1e4978413e0" containerName="mariadb-account-create" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016843 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="36757dfa-8de4-475b-bf51-7320d55c16c7" containerName="mariadb-account-create" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016854 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" containerName="glance-httpd" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.016871 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb63537-19e3-4a3c-a2b2-2b708c79fded" containerName="init" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.017837 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.021557 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.022554 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.024102 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.087269 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfb2j\" (UniqueName: \"kubernetes.io/projected/a8fbeb79-b731-40fe-867d-1df6c03db681-kube-api-access-pfb2j\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.087336 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8fbeb79-b731-40fe-867d-1df6c03db681-logs\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.087404 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.087513 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.087643 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.087688 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.087713 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.087744 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a8fbeb79-b731-40fe-867d-1df6c03db681-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.148354 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.191744 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-logs\") pod \"09623c61-6d40-4dd4-8f66-e9b398e9747f\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.191887 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-httpd-run\") pod \"09623c61-6d40-4dd4-8f66-e9b398e9747f\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.191935 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-config-data\") pod \"09623c61-6d40-4dd4-8f66-e9b398e9747f\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.191962 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm6jw\" (UniqueName: \"kubernetes.io/projected/09623c61-6d40-4dd4-8f66-e9b398e9747f-kube-api-access-xm6jw\") pod \"09623c61-6d40-4dd4-8f66-e9b398e9747f\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192186 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"09623c61-6d40-4dd4-8f66-e9b398e9747f\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192230 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-scripts\") pod \"09623c61-6d40-4dd4-8f66-e9b398e9747f\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192259 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-combined-ca-bundle\") pod \"09623c61-6d40-4dd4-8f66-e9b398e9747f\" (UID: \"09623c61-6d40-4dd4-8f66-e9b398e9747f\") " Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192628 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192695 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192718 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192754 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a8fbeb79-b731-40fe-867d-1df6c03db681-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192853 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfb2j\" (UniqueName: \"kubernetes.io/projected/a8fbeb79-b731-40fe-867d-1df6c03db681-kube-api-access-pfb2j\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192891 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8fbeb79-b731-40fe-867d-1df6c03db681-logs\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.192976 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.193103 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.198115 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a8fbeb79-b731-40fe-867d-1df6c03db681-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.198596 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.198725 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.211004 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.211475 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-logs" (OuterVolumeSpecName: "logs") pod "09623c61-6d40-4dd4-8f66-e9b398e9747f" (UID: "09623c61-6d40-4dd4-8f66-e9b398e9747f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.211961 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "09623c61-6d40-4dd4-8f66-e9b398e9747f" (UID: "09623c61-6d40-4dd4-8f66-e9b398e9747f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.198606 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8fbeb79-b731-40fe-867d-1df6c03db681-logs\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.215664 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.215748 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8fbeb79-b731-40fe-867d-1df6c03db681-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.217703 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09623c61-6d40-4dd4-8f66-e9b398e9747f-kube-api-access-xm6jw" (OuterVolumeSpecName: "kube-api-access-xm6jw") pod "09623c61-6d40-4dd4-8f66-e9b398e9747f" (UID: "09623c61-6d40-4dd4-8f66-e9b398e9747f"). InnerVolumeSpecName "kube-api-access-xm6jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.220762 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7dcb5d4fcc-q9spz"] Oct 08 14:23:21 crc kubenswrapper[4789]: E1008 14:23:21.221921 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerName="glance-log" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.221941 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerName="glance-log" Oct 08 14:23:21 crc kubenswrapper[4789]: E1008 14:23:21.221963 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerName="glance-httpd" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.221971 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerName="glance-httpd" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.222378 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerName="glance-log" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.222408 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="09623c61-6d40-4dd4-8f66-e9b398e9747f" containerName="glance-httpd" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.223731 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.228327 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.228409 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.233181 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "09623c61-6d40-4dd4-8f66-e9b398e9747f" (UID: "09623c61-6d40-4dd4-8f66-e9b398e9747f"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.242825 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7dcb5d4fcc-q9spz"] Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.243115 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfb2j\" (UniqueName: \"kubernetes.io/projected/a8fbeb79-b731-40fe-867d-1df6c03db681-kube-api-access-pfb2j\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.263201 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-scripts" (OuterVolumeSpecName: "scripts") pod "09623c61-6d40-4dd4-8f66-e9b398e9747f" (UID: "09623c61-6d40-4dd4-8f66-e9b398e9747f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296146 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-config\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296201 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-ovndb-tls-certs\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296229 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wln69\" (UniqueName: \"kubernetes.io/projected/1532f955-2e73-4a9a-b76a-507c2c545db0-kube-api-access-wln69\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296261 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-internal-tls-certs\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296312 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-public-tls-certs\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296369 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-httpd-config\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296388 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-combined-ca-bundle\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296452 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296462 4789 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09623c61-6d40-4dd4-8f66-e9b398e9747f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296471 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm6jw\" (UniqueName: \"kubernetes.io/projected/09623c61-6d40-4dd4-8f66-e9b398e9747f-kube-api-access-xm6jw\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296490 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.296500 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.297649 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"a8fbeb79-b731-40fe-867d-1df6c03db681\") " pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.321674 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09623c61-6d40-4dd4-8f66-e9b398e9747f" (UID: "09623c61-6d40-4dd4-8f66-e9b398e9747f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.330120 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.357334 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-config-data" (OuterVolumeSpecName: "config-data") pod "09623c61-6d40-4dd4-8f66-e9b398e9747f" (UID: "09623c61-6d40-4dd4-8f66-e9b398e9747f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.397781 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-config\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.398063 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-ovndb-tls-certs\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.398208 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wln69\" (UniqueName: \"kubernetes.io/projected/1532f955-2e73-4a9a-b76a-507c2c545db0-kube-api-access-wln69\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.398337 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-internal-tls-certs\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.400221 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-public-tls-certs\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.400538 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-httpd-config\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.400655 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-combined-ca-bundle\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.404451 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.404662 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.404759 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09623c61-6d40-4dd4-8f66-e9b398e9747f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.410614 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-config\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.410630 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-ovndb-tls-certs\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.411235 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-internal-tls-certs\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.420058 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-combined-ca-bundle\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.420434 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-httpd-config\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.420627 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wln69\" (UniqueName: \"kubernetes.io/projected/1532f955-2e73-4a9a-b76a-507c2c545db0-kube-api-access-wln69\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.422090 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1532f955-2e73-4a9a-b76a-507c2c545db0-public-tls-certs\") pod \"neutron-7dcb5d4fcc-q9spz\" (UID: \"1532f955-2e73-4a9a-b76a-507c2c545db0\") " pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.440724 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.443416 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.667944 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"09623c61-6d40-4dd4-8f66-e9b398e9747f","Type":"ContainerDied","Data":"09dc8a0cf06f3f2d85816d0ab99fe3fda0c1d776690379972cd6f17ea7b577a4"} Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.668052 4789 scope.go:117] "RemoveContainer" containerID="4f2de78d9e2f4d0f6eafd577138bee33aa364de14faf849c43cca8cd6eb3e1d9" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.668093 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.706924 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.724920 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.744845 4789 scope.go:117] "RemoveContainer" containerID="b3542835961fe21f75337e37f743e8b812c03c6d3ea91479e29f928f4774b96f" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.749054 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.750862 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.754792 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.755038 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.769899 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.813401 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-462gn\" (UniqueName: \"kubernetes.io/projected/1fbecfec-ecd2-458e-b04f-e715a9ec4186-kube-api-access-462gn\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.813445 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-config-data\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.813486 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1fbecfec-ecd2-458e-b04f-e715a9ec4186-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.814513 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.814705 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.815084 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.815175 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-scripts\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.815237 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fbecfec-ecd2-458e-b04f-e715a9ec4186-logs\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.917798 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-462gn\" (UniqueName: \"kubernetes.io/projected/1fbecfec-ecd2-458e-b04f-e715a9ec4186-kube-api-access-462gn\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.917847 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-config-data\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.917882 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1fbecfec-ecd2-458e-b04f-e715a9ec4186-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.917970 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.918021 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.918051 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.918084 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-scripts\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.918113 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fbecfec-ecd2-458e-b04f-e715a9ec4186-logs\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.918511 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1fbecfec-ecd2-458e-b04f-e715a9ec4186-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.918585 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.919141 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fbecfec-ecd2-458e-b04f-e715a9ec4186-logs\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.925970 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.928259 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-scripts\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.946259 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-config-data\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.949451 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-462gn\" (UniqueName: \"kubernetes.io/projected/1fbecfec-ecd2-458e-b04f-e715a9ec4186-kube-api-access-462gn\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.962614 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fbecfec-ecd2-458e-b04f-e715a9ec4186-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:21 crc kubenswrapper[4789]: I1008 14:23:21.989130 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"1fbecfec-ecd2-458e-b04f-e715a9ec4186\") " pod="openstack/glance-default-external-api-0" Oct 08 14:23:22 crc kubenswrapper[4789]: I1008 14:23:22.081704 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 14:23:22 crc kubenswrapper[4789]: I1008 14:23:22.106114 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 14:23:22 crc kubenswrapper[4789]: I1008 14:23:22.216566 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7dcb5d4fcc-q9spz"] Oct 08 14:23:22 crc kubenswrapper[4789]: I1008 14:23:22.697122 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a8fbeb79-b731-40fe-867d-1df6c03db681","Type":"ContainerStarted","Data":"79f87888ee907a3f9962df77cfac01656b2c56d2586c7e184452e6e086e0355d"} Oct 08 14:23:22 crc kubenswrapper[4789]: I1008 14:23:22.699133 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7dcb5d4fcc-q9spz" event={"ID":"1532f955-2e73-4a9a-b76a-507c2c545db0","Type":"ContainerStarted","Data":"568dd985f3e9f45520825bd4fb0660e9d744c26edf2a2ebf9e499b547fc0ccff"} Oct 08 14:23:22 crc kubenswrapper[4789]: I1008 14:23:22.746936 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09623c61-6d40-4dd4-8f66-e9b398e9747f" path="/var/lib/kubelet/pods/09623c61-6d40-4dd4-8f66-e9b398e9747f/volumes" Oct 08 14:23:22 crc kubenswrapper[4789]: I1008 14:23:22.750276 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d13191a-daae-49b6-aabe-d9c0ab0018bf" path="/var/lib/kubelet/pods/1d13191a-daae-49b6-aabe-d9c0ab0018bf/volumes" Oct 08 14:23:22 crc kubenswrapper[4789]: I1008 14:23:22.763454 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 14:23:23 crc kubenswrapper[4789]: I1008 14:23:23.725372 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a8fbeb79-b731-40fe-867d-1df6c03db681","Type":"ContainerStarted","Data":"16322965de35ef5fe82fca9ed76482489102f12f6aa0608c93a1490f32927d8b"} Oct 08 14:23:23 crc kubenswrapper[4789]: I1008 14:23:23.742959 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7dcb5d4fcc-q9spz" event={"ID":"1532f955-2e73-4a9a-b76a-507c2c545db0","Type":"ContainerStarted","Data":"34ef66cbefed3ae6d2bc23b77d358e92b575def7bbd5e1ef70fa014102dd6819"} Oct 08 14:23:23 crc kubenswrapper[4789]: I1008 14:23:23.743020 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7dcb5d4fcc-q9spz" event={"ID":"1532f955-2e73-4a9a-b76a-507c2c545db0","Type":"ContainerStarted","Data":"fa447e538bbbbd96b876131c0d33b595b9a0ecc5eb8fc5db79898c6260c3e6e9"} Oct 08 14:23:23 crc kubenswrapper[4789]: I1008 14:23:23.743499 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:23 crc kubenswrapper[4789]: I1008 14:23:23.755934 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1fbecfec-ecd2-458e-b04f-e715a9ec4186","Type":"ContainerStarted","Data":"789e49f5c59c0cbdae577e501f53d9096c1ab9c6bbb56234dc051ad75aabbbb6"} Oct 08 14:23:23 crc kubenswrapper[4789]: I1008 14:23:23.756060 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1fbecfec-ecd2-458e-b04f-e715a9ec4186","Type":"ContainerStarted","Data":"2616cd7042be42a51977bb83d5a0b6c60f875bf70827e7c8dcae178281652075"} Oct 08 14:23:23 crc kubenswrapper[4789]: I1008 14:23:23.779370 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7dcb5d4fcc-q9spz" podStartSLOduration=2.779351853 podStartE2EDuration="2.779351853s" podCreationTimestamp="2025-10-08 14:23:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:23.766509495 +0000 UTC m=+1343.673256987" watchObservedRunningTime="2025-10-08 14:23:23.779351853 +0000 UTC m=+1343.686099345" Oct 08 14:23:24 crc kubenswrapper[4789]: I1008 14:23:24.767134 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a8fbeb79-b731-40fe-867d-1df6c03db681","Type":"ContainerStarted","Data":"6512c0976e14730977f839e2612f5a841bea9aee52f5969cef6ce39330a3cd2f"} Oct 08 14:23:24 crc kubenswrapper[4789]: I1008 14:23:24.769107 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1fbecfec-ecd2-458e-b04f-e715a9ec4186","Type":"ContainerStarted","Data":"0e5cf598faa60e628b19b5ada12a4cc260ae6c3960520b1bdcd8ed5b66826cc3"} Oct 08 14:23:24 crc kubenswrapper[4789]: I1008 14:23:24.793092 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.793073472 podStartE2EDuration="4.793073472s" podCreationTimestamp="2025-10-08 14:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:24.787582623 +0000 UTC m=+1344.694330115" watchObservedRunningTime="2025-10-08 14:23:24.793073472 +0000 UTC m=+1344.699820964" Oct 08 14:23:24 crc kubenswrapper[4789]: I1008 14:23:24.813850 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.813830544 podStartE2EDuration="3.813830544s" podCreationTimestamp="2025-10-08 14:23:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:24.80959732 +0000 UTC m=+1344.716344822" watchObservedRunningTime="2025-10-08 14:23:24.813830544 +0000 UTC m=+1344.720578036" Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.139230 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-a3a4-account-create-7wxzg"] Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.140920 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a3a4-account-create-7wxzg" Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.145526 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.158620 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a3a4-account-create-7wxzg"] Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.208181 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54qrm\" (UniqueName: \"kubernetes.io/projected/bbc3a435-e2f9-4353-910c-b62c505278d7-kube-api-access-54qrm\") pod \"nova-cell0-a3a4-account-create-7wxzg\" (UID: \"bbc3a435-e2f9-4353-910c-b62c505278d7\") " pod="openstack/nova-cell0-a3a4-account-create-7wxzg" Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.310290 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54qrm\" (UniqueName: \"kubernetes.io/projected/bbc3a435-e2f9-4353-910c-b62c505278d7-kube-api-access-54qrm\") pod \"nova-cell0-a3a4-account-create-7wxzg\" (UID: \"bbc3a435-e2f9-4353-910c-b62c505278d7\") " pod="openstack/nova-cell0-a3a4-account-create-7wxzg" Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.327866 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.347468 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54qrm\" (UniqueName: \"kubernetes.io/projected/bbc3a435-e2f9-4353-910c-b62c505278d7-kube-api-access-54qrm\") pod \"nova-cell0-a3a4-account-create-7wxzg\" (UID: \"bbc3a435-e2f9-4353-910c-b62c505278d7\") " pod="openstack/nova-cell0-a3a4-account-create-7wxzg" Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.416482 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9bf94f7c-84pd2"] Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.416737 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" podUID="632b4bda-256d-4237-aade-b2cdffbf2d93" containerName="dnsmasq-dns" containerID="cri-o://cd96ed81cb45e4420f241225df836cd12fce3b507ef2cf97afc720a3e2df01a7" gracePeriod=10 Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.461181 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a3a4-account-create-7wxzg" Oct 08 14:23:25 crc kubenswrapper[4789]: W1008 14:23:25.484099 4789 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfb63537_19e3_4a3c_a2b2_2b708c79fded.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfb63537_19e3_4a3c_a2b2_2b708c79fded.slice: no such file or directory Oct 08 14:23:25 crc kubenswrapper[4789]: W1008 14:23:25.484755 4789 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09623c61_6d40_4dd4_8f66_e9b398e9747f.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09623c61_6d40_4dd4_8f66_e9b398e9747f.slice: no such file or directory Oct 08 14:23:25 crc kubenswrapper[4789]: W1008 14:23:25.487099 4789 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36757dfa_8de4_475b_bf51_7320d55c16c7.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36757dfa_8de4_475b_bf51_7320d55c16c7.slice: no such file or directory Oct 08 14:23:25 crc kubenswrapper[4789]: W1008 14:23:25.487170 4789 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d13191a_daae_49b6_aabe_d9c0ab0018bf.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d13191a_daae_49b6_aabe_d9c0ab0018bf.slice: no such file or directory Oct 08 14:23:25 crc kubenswrapper[4789]: W1008 14:23:25.487208 4789 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b985f8e_1745_4c46_9f99_d1e4978413e0.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b985f8e_1745_4c46_9f99_d1e4978413e0.slice: no such file or directory Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.780271 4789 generic.go:334] "Generic (PLEG): container finished" podID="632b4bda-256d-4237-aade-b2cdffbf2d93" containerID="cd96ed81cb45e4420f241225df836cd12fce3b507ef2cf97afc720a3e2df01a7" exitCode=0 Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.780342 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" event={"ID":"632b4bda-256d-4237-aade-b2cdffbf2d93","Type":"ContainerDied","Data":"cd96ed81cb45e4420f241225df836cd12fce3b507ef2cf97afc720a3e2df01a7"} Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.784859 4789 generic.go:334] "Generic (PLEG): container finished" podID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerID="da4dcaa236c25382f0333cf0697d930eadaf476ef72aaa894d3d401749065420" exitCode=137 Oct 08 14:23:25 crc kubenswrapper[4789]: I1008 14:23:25.785884 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerDied","Data":"da4dcaa236c25382f0333cf0697d930eadaf476ef72aaa894d3d401749065420"} Oct 08 14:23:25 crc kubenswrapper[4789]: E1008 14:23:25.798398 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod632b4bda_256d_4237_aade_b2cdffbf2d93.slice/crio-cd96ed81cb45e4420f241225df836cd12fce3b507ef2cf97afc720a3e2df01a7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod632b4bda_256d_4237_aade_b2cdffbf2d93.slice/crio-conmon-cd96ed81cb45e4420f241225df836cd12fce3b507ef2cf97afc720a3e2df01a7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod835cb27d_9cce_484d_95bd_ee88abf5a2dc.slice/crio-da4dcaa236c25382f0333cf0697d930eadaf476ef72aaa894d3d401749065420.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod835cb27d_9cce_484d_95bd_ee88abf5a2dc.slice/crio-conmon-da4dcaa236c25382f0333cf0697d930eadaf476ef72aaa894d3d401749065420.scope\": RecentStats: unable to find data in memory cache]" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.106702 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.162584 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236250 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-run-httpd\") pod \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236318 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-log-httpd\") pod \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236598 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb7hx\" (UniqueName: \"kubernetes.io/projected/835cb27d-9cce-484d-95bd-ee88abf5a2dc-kube-api-access-xb7hx\") pod \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236732 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-config-data\") pod \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236768 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsdst\" (UniqueName: \"kubernetes.io/projected/632b4bda-256d-4237-aade-b2cdffbf2d93-kube-api-access-tsdst\") pod \"632b4bda-256d-4237-aade-b2cdffbf2d93\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236811 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-swift-storage-0\") pod \"632b4bda-256d-4237-aade-b2cdffbf2d93\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236841 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-scripts\") pod \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236861 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-config\") pod \"632b4bda-256d-4237-aade-b2cdffbf2d93\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236889 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-sb\") pod \"632b4bda-256d-4237-aade-b2cdffbf2d93\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.236934 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-combined-ca-bundle\") pod \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.237073 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-svc\") pod \"632b4bda-256d-4237-aade-b2cdffbf2d93\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.237104 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-nb\") pod \"632b4bda-256d-4237-aade-b2cdffbf2d93\" (UID: \"632b4bda-256d-4237-aade-b2cdffbf2d93\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.237161 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-sg-core-conf-yaml\") pod \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\" (UID: \"835cb27d-9cce-484d-95bd-ee88abf5a2dc\") " Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.238357 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "835cb27d-9cce-484d-95bd-ee88abf5a2dc" (UID: "835cb27d-9cce-484d-95bd-ee88abf5a2dc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.238660 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "835cb27d-9cce-484d-95bd-ee88abf5a2dc" (UID: "835cb27d-9cce-484d-95bd-ee88abf5a2dc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.247870 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/835cb27d-9cce-484d-95bd-ee88abf5a2dc-kube-api-access-xb7hx" (OuterVolumeSpecName: "kube-api-access-xb7hx") pod "835cb27d-9cce-484d-95bd-ee88abf5a2dc" (UID: "835cb27d-9cce-484d-95bd-ee88abf5a2dc"). InnerVolumeSpecName "kube-api-access-xb7hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.251083 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/632b4bda-256d-4237-aade-b2cdffbf2d93-kube-api-access-tsdst" (OuterVolumeSpecName: "kube-api-access-tsdst") pod "632b4bda-256d-4237-aade-b2cdffbf2d93" (UID: "632b4bda-256d-4237-aade-b2cdffbf2d93"). InnerVolumeSpecName "kube-api-access-tsdst". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.262669 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a3a4-account-create-7wxzg"] Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.264131 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-scripts" (OuterVolumeSpecName: "scripts") pod "835cb27d-9cce-484d-95bd-ee88abf5a2dc" (UID: "835cb27d-9cce-484d-95bd-ee88abf5a2dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.297414 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "835cb27d-9cce-484d-95bd-ee88abf5a2dc" (UID: "835cb27d-9cce-484d-95bd-ee88abf5a2dc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.342804 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.343531 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.343559 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/835cb27d-9cce-484d-95bd-ee88abf5a2dc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.343599 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb7hx\" (UniqueName: \"kubernetes.io/projected/835cb27d-9cce-484d-95bd-ee88abf5a2dc-kube-api-access-xb7hx\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.343619 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsdst\" (UniqueName: \"kubernetes.io/projected/632b4bda-256d-4237-aade-b2cdffbf2d93-kube-api-access-tsdst\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.343712 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.356802 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "632b4bda-256d-4237-aade-b2cdffbf2d93" (UID: "632b4bda-256d-4237-aade-b2cdffbf2d93"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.373365 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "632b4bda-256d-4237-aade-b2cdffbf2d93" (UID: "632b4bda-256d-4237-aade-b2cdffbf2d93"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.384270 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "632b4bda-256d-4237-aade-b2cdffbf2d93" (UID: "632b4bda-256d-4237-aade-b2cdffbf2d93"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.404621 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-config" (OuterVolumeSpecName: "config") pod "632b4bda-256d-4237-aade-b2cdffbf2d93" (UID: "632b4bda-256d-4237-aade-b2cdffbf2d93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.414571 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "632b4bda-256d-4237-aade-b2cdffbf2d93" (UID: "632b4bda-256d-4237-aade-b2cdffbf2d93"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.414709 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-config-data" (OuterVolumeSpecName: "config-data") pod "835cb27d-9cce-484d-95bd-ee88abf5a2dc" (UID: "835cb27d-9cce-484d-95bd-ee88abf5a2dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.432168 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "835cb27d-9cce-484d-95bd-ee88abf5a2dc" (UID: "835cb27d-9cce-484d-95bd-ee88abf5a2dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.434167 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.434323 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.445747 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.445973 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.446063 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.446129 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.446208 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/835cb27d-9cce-484d-95bd-ee88abf5a2dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.446278 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.446349 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/632b4bda-256d-4237-aade-b2cdffbf2d93-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.799396 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.799401 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bf94f7c-84pd2" event={"ID":"632b4bda-256d-4237-aade-b2cdffbf2d93","Type":"ContainerDied","Data":"ab1c0ce8619a6e3796813ec892fda7994e5ae4c54d30dc4d000ce95030eadcb4"} Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.799931 4789 scope.go:117] "RemoveContainer" containerID="cd96ed81cb45e4420f241225df836cd12fce3b507ef2cf97afc720a3e2df01a7" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.803561 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a3a4-account-create-7wxzg" event={"ID":"bbc3a435-e2f9-4353-910c-b62c505278d7","Type":"ContainerStarted","Data":"7e200e25890c4ed28a8bd1c4effa8ac23aa8fce4567094f4e1a571ac33a876df"} Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.819694 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"835cb27d-9cce-484d-95bd-ee88abf5a2dc","Type":"ContainerDied","Data":"6d8b42937a4b740b7857685b32e1387f28238f8bdf6ad79dfdf0cd9375163261"} Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.819758 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.832014 4789 scope.go:117] "RemoveContainer" containerID="1efbe07b43550f1d5907871e1eefe02ab9ed15279288204ed952b44c292548ef" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.840677 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9bf94f7c-84pd2"] Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.857063 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f9bf94f7c-84pd2"] Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.865329 4789 scope.go:117] "RemoveContainer" containerID="d9686918aa16832170c166e4e73c2e45e5490b24d9f856084bdc0e6b31728181" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.874955 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.885106 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.895124 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:23:26 crc kubenswrapper[4789]: E1008 14:23:26.895696 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="proxy-httpd" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.895719 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="proxy-httpd" Oct 08 14:23:26 crc kubenswrapper[4789]: E1008 14:23:26.895740 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="ceilometer-central-agent" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.895749 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="ceilometer-central-agent" Oct 08 14:23:26 crc kubenswrapper[4789]: E1008 14:23:26.895766 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632b4bda-256d-4237-aade-b2cdffbf2d93" containerName="dnsmasq-dns" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.895774 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="632b4bda-256d-4237-aade-b2cdffbf2d93" containerName="dnsmasq-dns" Oct 08 14:23:26 crc kubenswrapper[4789]: E1008 14:23:26.895795 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="ceilometer-notification-agent" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.895803 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="ceilometer-notification-agent" Oct 08 14:23:26 crc kubenswrapper[4789]: E1008 14:23:26.895822 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="sg-core" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.895829 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="sg-core" Oct 08 14:23:26 crc kubenswrapper[4789]: E1008 14:23:26.895843 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632b4bda-256d-4237-aade-b2cdffbf2d93" containerName="init" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.895851 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="632b4bda-256d-4237-aade-b2cdffbf2d93" containerName="init" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.896099 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="632b4bda-256d-4237-aade-b2cdffbf2d93" containerName="dnsmasq-dns" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.896129 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="ceilometer-central-agent" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.896148 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="ceilometer-notification-agent" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.896162 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="proxy-httpd" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.896181 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" containerName="sg-core" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.898733 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.901658 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.902655 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.903268 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.956145 4789 scope.go:117] "RemoveContainer" containerID="2a0ef0e58fcc4cb78ad4e12987bc81057e0bfe689a2a190455ee1e42efc2f5dd" Oct 08 14:23:26 crc kubenswrapper[4789]: I1008 14:23:26.986387 4789 scope.go:117] "RemoveContainer" containerID="da4dcaa236c25382f0333cf0697d930eadaf476ef72aaa894d3d401749065420" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.031462 4789 scope.go:117] "RemoveContainer" containerID="f66f2fc2457dc225f1d54437ce02e187a00ef23ef69788900141659196068692" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.057821 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-config-data\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.058249 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-run-httpd\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.058397 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-log-httpd\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.058524 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.058685 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-scripts\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.058774 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.058878 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w4dg\" (UniqueName: \"kubernetes.io/projected/fa4caa03-96c7-4142-a059-8f331cbbf822-kube-api-access-8w4dg\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.160273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-config-data\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.160374 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-run-httpd\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.160399 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-log-httpd\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.160420 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.160476 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-scripts\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.160496 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.160522 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w4dg\" (UniqueName: \"kubernetes.io/projected/fa4caa03-96c7-4142-a059-8f331cbbf822-kube-api-access-8w4dg\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.161157 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-log-httpd\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.161266 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-run-httpd\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.167331 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-scripts\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.172901 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.173544 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-config-data\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.177968 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w4dg\" (UniqueName: \"kubernetes.io/projected/fa4caa03-96c7-4142-a059-8f331cbbf822-kube-api-access-8w4dg\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.178942 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.221623 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.675375 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.830882 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a3a4-account-create-7wxzg" event={"ID":"bbc3a435-e2f9-4353-910c-b62c505278d7","Type":"ContainerStarted","Data":"4af4d007406050d269006b2dfb41b07b88dab2589dd6f8ac0cb2383e58f5d12e"} Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.833778 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerStarted","Data":"11cb66d15cd10ee7388bfa73962fc056ca738aa7b6073194b33ea9d700711052"} Oct 08 14:23:27 crc kubenswrapper[4789]: I1008 14:23:27.845975 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-a3a4-account-create-7wxzg" podStartSLOduration=2.845957856 podStartE2EDuration="2.845957856s" podCreationTimestamp="2025-10-08 14:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:27.843928251 +0000 UTC m=+1347.750675743" watchObservedRunningTime="2025-10-08 14:23:27.845957856 +0000 UTC m=+1347.752705348" Oct 08 14:23:28 crc kubenswrapper[4789]: I1008 14:23:28.742871 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="632b4bda-256d-4237-aade-b2cdffbf2d93" path="/var/lib/kubelet/pods/632b4bda-256d-4237-aade-b2cdffbf2d93/volumes" Oct 08 14:23:28 crc kubenswrapper[4789]: I1008 14:23:28.744292 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="835cb27d-9cce-484d-95bd-ee88abf5a2dc" path="/var/lib/kubelet/pods/835cb27d-9cce-484d-95bd-ee88abf5a2dc/volumes" Oct 08 14:23:28 crc kubenswrapper[4789]: I1008 14:23:28.873164 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerStarted","Data":"4bef4bad303b0844f63c432c915c5fcbfaab91aa4c11130e7915d8fc8db26b61"} Oct 08 14:23:28 crc kubenswrapper[4789]: I1008 14:23:28.875294 4789 generic.go:334] "Generic (PLEG): container finished" podID="bbc3a435-e2f9-4353-910c-b62c505278d7" containerID="4af4d007406050d269006b2dfb41b07b88dab2589dd6f8ac0cb2383e58f5d12e" exitCode=0 Oct 08 14:23:28 crc kubenswrapper[4789]: I1008 14:23:28.875336 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a3a4-account-create-7wxzg" event={"ID":"bbc3a435-e2f9-4353-910c-b62c505278d7","Type":"ContainerDied","Data":"4af4d007406050d269006b2dfb41b07b88dab2589dd6f8ac0cb2383e58f5d12e"} Oct 08 14:23:29 crc kubenswrapper[4789]: I1008 14:23:29.887384 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerStarted","Data":"9e91c3c4bf79c57d9547837c1ba31c3f0074f731b9965477c64868c2b0a6e30f"} Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.496820 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a3a4-account-create-7wxzg" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.629849 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54qrm\" (UniqueName: \"kubernetes.io/projected/bbc3a435-e2f9-4353-910c-b62c505278d7-kube-api-access-54qrm\") pod \"bbc3a435-e2f9-4353-910c-b62c505278d7\" (UID: \"bbc3a435-e2f9-4353-910c-b62c505278d7\") " Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.636864 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbc3a435-e2f9-4353-910c-b62c505278d7-kube-api-access-54qrm" (OuterVolumeSpecName: "kube-api-access-54qrm") pod "bbc3a435-e2f9-4353-910c-b62c505278d7" (UID: "bbc3a435-e2f9-4353-910c-b62c505278d7"). InnerVolumeSpecName "kube-api-access-54qrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.693591 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.736778 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54qrm\" (UniqueName: \"kubernetes.io/projected/bbc3a435-e2f9-4353-910c-b62c505278d7-kube-api-access-54qrm\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.837957 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-etc-machine-id\") pod \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.838028 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data-custom\") pod \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.838077 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data\") pod \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.838076 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "057a92ec-1b02-446f-bc0e-3ac7d2344d2c" (UID: "057a92ec-1b02-446f-bc0e-3ac7d2344d2c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.838110 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb5kk\" (UniqueName: \"kubernetes.io/projected/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-kube-api-access-rb5kk\") pod \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.838152 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-scripts\") pod \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.838200 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-combined-ca-bundle\") pod \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.838329 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-logs\") pod \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\" (UID: \"057a92ec-1b02-446f-bc0e-3ac7d2344d2c\") " Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.838731 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-logs" (OuterVolumeSpecName: "logs") pod "057a92ec-1b02-446f-bc0e-3ac7d2344d2c" (UID: "057a92ec-1b02-446f-bc0e-3ac7d2344d2c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.839132 4789 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.839157 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.843808 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-kube-api-access-rb5kk" (OuterVolumeSpecName: "kube-api-access-rb5kk") pod "057a92ec-1b02-446f-bc0e-3ac7d2344d2c" (UID: "057a92ec-1b02-446f-bc0e-3ac7d2344d2c"). InnerVolumeSpecName "kube-api-access-rb5kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.849154 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "057a92ec-1b02-446f-bc0e-3ac7d2344d2c" (UID: "057a92ec-1b02-446f-bc0e-3ac7d2344d2c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.855132 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-scripts" (OuterVolumeSpecName: "scripts") pod "057a92ec-1b02-446f-bc0e-3ac7d2344d2c" (UID: "057a92ec-1b02-446f-bc0e-3ac7d2344d2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.907621 4789 generic.go:334] "Generic (PLEG): container finished" podID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerID="3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11" exitCode=137 Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.907712 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.907735 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"057a92ec-1b02-446f-bc0e-3ac7d2344d2c","Type":"ContainerDied","Data":"3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11"} Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.908213 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"057a92ec-1b02-446f-bc0e-3ac7d2344d2c","Type":"ContainerDied","Data":"3c780b9ae041f2f359253c6120ff30cdc55859fbdf89df434c9389ae35d1d30b"} Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.908245 4789 scope.go:117] "RemoveContainer" containerID="3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.915111 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "057a92ec-1b02-446f-bc0e-3ac7d2344d2c" (UID: "057a92ec-1b02-446f-bc0e-3ac7d2344d2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.918129 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerStarted","Data":"2e2c20f13cc3e8474e669c53397f2098e91c7a8e073298deada1d92d32bf3b46"} Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.921201 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a3a4-account-create-7wxzg" event={"ID":"bbc3a435-e2f9-4353-910c-b62c505278d7","Type":"ContainerDied","Data":"7e200e25890c4ed28a8bd1c4effa8ac23aa8fce4567094f4e1a571ac33a876df"} Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.921235 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e200e25890c4ed28a8bd1c4effa8ac23aa8fce4567094f4e1a571ac33a876df" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.921255 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a3a4-account-create-7wxzg" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.939169 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data" (OuterVolumeSpecName: "config-data") pod "057a92ec-1b02-446f-bc0e-3ac7d2344d2c" (UID: "057a92ec-1b02-446f-bc0e-3ac7d2344d2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.940674 4789 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.940713 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.940723 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb5kk\" (UniqueName: \"kubernetes.io/projected/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-kube-api-access-rb5kk\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.940734 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.940742 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a92ec-1b02-446f-bc0e-3ac7d2344d2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.947318 4789 scope.go:117] "RemoveContainer" containerID="e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.979687 4789 scope.go:117] "RemoveContainer" containerID="3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11" Oct 08 14:23:30 crc kubenswrapper[4789]: E1008 14:23:30.983389 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11\": container with ID starting with 3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11 not found: ID does not exist" containerID="3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.983432 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11"} err="failed to get container status \"3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11\": rpc error: code = NotFound desc = could not find container \"3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11\": container with ID starting with 3aaa093c06d6c4dc0bf09de6285b928a8a1e7b5c4baa701db8df7ff1d3d2fe11 not found: ID does not exist" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.983467 4789 scope.go:117] "RemoveContainer" containerID="e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49" Oct 08 14:23:30 crc kubenswrapper[4789]: E1008 14:23:30.985250 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49\": container with ID starting with e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49 not found: ID does not exist" containerID="e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49" Oct 08 14:23:30 crc kubenswrapper[4789]: I1008 14:23:30.985289 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49"} err="failed to get container status \"e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49\": rpc error: code = NotFound desc = could not find container \"e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49\": container with ID starting with e5622067175c6053426a02f7835b488be3e5169a0e79fdeb610f1ae978991c49 not found: ID does not exist" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.242907 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.297360 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.309687 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 14:23:31 crc kubenswrapper[4789]: E1008 14:23:31.310102 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerName="cinder-api" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.310125 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerName="cinder-api" Oct 08 14:23:31 crc kubenswrapper[4789]: E1008 14:23:31.310135 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbc3a435-e2f9-4353-910c-b62c505278d7" containerName="mariadb-account-create" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.310144 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbc3a435-e2f9-4353-910c-b62c505278d7" containerName="mariadb-account-create" Oct 08 14:23:31 crc kubenswrapper[4789]: E1008 14:23:31.310190 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerName="cinder-api-log" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.310199 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerName="cinder-api-log" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.310401 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerName="cinder-api" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.310428 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" containerName="cinder-api-log" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.310447 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbc3a435-e2f9-4353-910c-b62c505278d7" containerName="mariadb-account-create" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.318631 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.321395 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.321604 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.321735 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.327525 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.443640 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.443718 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.466566 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a72d40d2-86d8-406f-9f5c-adec376cc947-logs\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.466619 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-config-data\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.466641 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cld78\" (UniqueName: \"kubernetes.io/projected/a72d40d2-86d8-406f-9f5c-adec376cc947-kube-api-access-cld78\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.466668 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-config-data-custom\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.466684 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.466928 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-scripts\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.467135 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a72d40d2-86d8-406f-9f5c-adec376cc947-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.467272 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.467330 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.499051 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.524690 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.568677 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-config-data-custom\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.568724 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.568773 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-scripts\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.569009 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a72d40d2-86d8-406f-9f5c-adec376cc947-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.569066 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.569094 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.569160 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a72d40d2-86d8-406f-9f5c-adec376cc947-logs\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.569170 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a72d40d2-86d8-406f-9f5c-adec376cc947-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.569219 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-config-data\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.569242 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cld78\" (UniqueName: \"kubernetes.io/projected/a72d40d2-86d8-406f-9f5c-adec376cc947-kube-api-access-cld78\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.569568 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a72d40d2-86d8-406f-9f5c-adec376cc947-logs\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.573193 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.573267 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.575646 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.576330 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-config-data-custom\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.585786 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cld78\" (UniqueName: \"kubernetes.io/projected/a72d40d2-86d8-406f-9f5c-adec376cc947-kube-api-access-cld78\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.588423 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-scripts\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.592287 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a72d40d2-86d8-406f-9f5c-adec376cc947-config-data\") pod \"cinder-api-0\" (UID: \"a72d40d2-86d8-406f-9f5c-adec376cc947\") " pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.639783 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.939924 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:31 crc kubenswrapper[4789]: I1008 14:23:31.940279 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.082967 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.083027 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.142375 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.161340 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.237858 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 14:23:32 crc kubenswrapper[4789]: W1008 14:23:32.256224 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda72d40d2_86d8_406f_9f5c_adec376cc947.slice/crio-dcbf1c37b96765d1038192ddf4462bfc8bd298394bbb3e88c92e69c714d72134 WatchSource:0}: Error finding container dcbf1c37b96765d1038192ddf4462bfc8bd298394bbb3e88c92e69c714d72134: Status 404 returned error can't find the container with id dcbf1c37b96765d1038192ddf4462bfc8bd298394bbb3e88c92e69c714d72134 Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.745848 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057a92ec-1b02-446f-bc0e-3ac7d2344d2c" path="/var/lib/kubelet/pods/057a92ec-1b02-446f-bc0e-3ac7d2344d2c/volumes" Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.949481 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a72d40d2-86d8-406f-9f5c-adec376cc947","Type":"ContainerStarted","Data":"dcbf1c37b96765d1038192ddf4462bfc8bd298394bbb3e88c92e69c714d72134"} Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.952463 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerStarted","Data":"a2909f922794801e50a3e6cbd2fd155131e523d6e101ea710c21f5f5af8d2e28"} Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.953097 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.953558 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 14:23:32 crc kubenswrapper[4789]: I1008 14:23:32.973070 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.933479386 podStartE2EDuration="6.973051686s" podCreationTimestamp="2025-10-08 14:23:26 +0000 UTC" firstStartedPulling="2025-10-08 14:23:27.680793781 +0000 UTC m=+1347.587541273" lastFinishedPulling="2025-10-08 14:23:31.720366081 +0000 UTC m=+1351.627113573" observedRunningTime="2025-10-08 14:23:32.969637153 +0000 UTC m=+1352.876384675" watchObservedRunningTime="2025-10-08 14:23:32.973051686 +0000 UTC m=+1352.879799188" Oct 08 14:23:33 crc kubenswrapper[4789]: I1008 14:23:33.963777 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a72d40d2-86d8-406f-9f5c-adec376cc947","Type":"ContainerStarted","Data":"c7edf4ce09d7acad09d2fa356be0a963e2044c503c4a6150b82a4d215d641f38"} Oct 08 14:23:33 crc kubenswrapper[4789]: I1008 14:23:33.964575 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 14:23:33 crc kubenswrapper[4789]: I1008 14:23:33.964617 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:23:33 crc kubenswrapper[4789]: I1008 14:23:33.964624 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:23:34 crc kubenswrapper[4789]: I1008 14:23:34.977079 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:23:34 crc kubenswrapper[4789]: I1008 14:23:34.977351 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:23:34 crc kubenswrapper[4789]: I1008 14:23:34.977063 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a72d40d2-86d8-406f-9f5c-adec376cc947","Type":"ContainerStarted","Data":"9ef90847892a93cc72b8763e79a938e9d2bf9c21389dd966635e3dd982d37ec7"} Oct 08 14:23:34 crc kubenswrapper[4789]: I1008 14:23:34.978272 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.003310 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.003288969 podStartE2EDuration="4.003288969s" podCreationTimestamp="2025-10-08 14:23:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:34.994731297 +0000 UTC m=+1354.901478789" watchObservedRunningTime="2025-10-08 14:23:35.003288969 +0000 UTC m=+1354.910036461" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.324838 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-75hcm"] Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.326305 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.343296 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-75hcm"] Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.344303 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-scripts\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.344354 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.344497 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-config-data\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.344589 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs77f\" (UniqueName: \"kubernetes.io/projected/c4779c61-4a62-4853-b877-fbf588daa8e8-kube-api-access-gs77f\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.345210 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.345414 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wwghs" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.345601 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.446505 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-scripts\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.446566 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.447052 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-config-data\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.447221 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs77f\" (UniqueName: \"kubernetes.io/projected/c4779c61-4a62-4853-b877-fbf588daa8e8-kube-api-access-gs77f\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.455831 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.459637 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-config-data\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.461378 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-scripts\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.477517 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs77f\" (UniqueName: \"kubernetes.io/projected/c4779c61-4a62-4853-b877-fbf588daa8e8-kube-api-access-gs77f\") pod \"nova-cell0-conductor-db-sync-75hcm\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:35 crc kubenswrapper[4789]: I1008 14:23:35.649751 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:23:36 crc kubenswrapper[4789]: W1008 14:23:36.141545 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4779c61_4a62_4853_b877_fbf588daa8e8.slice/crio-08cf6551f91428b7c627c96d8e8a3439f32ac915f3a65d321d557ccc33cbc16d WatchSource:0}: Error finding container 08cf6551f91428b7c627c96d8e8a3439f32ac915f3a65d321d557ccc33cbc16d: Status 404 returned error can't find the container with id 08cf6551f91428b7c627c96d8e8a3439f32ac915f3a65d321d557ccc33cbc16d Oct 08 14:23:36 crc kubenswrapper[4789]: I1008 14:23:36.148005 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-75hcm"] Oct 08 14:23:36 crc kubenswrapper[4789]: I1008 14:23:36.986918 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 14:23:36 crc kubenswrapper[4789]: I1008 14:23:36.987431 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:23:36 crc kubenswrapper[4789]: I1008 14:23:36.988250 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:36 crc kubenswrapper[4789]: I1008 14:23:36.988377 4789 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 14:23:36 crc kubenswrapper[4789]: I1008 14:23:36.994508 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 14:23:37 crc kubenswrapper[4789]: I1008 14:23:37.002532 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 14:23:37 crc kubenswrapper[4789]: I1008 14:23:37.003438 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-75hcm" event={"ID":"c4779c61-4a62-4853-b877-fbf588daa8e8","Type":"ContainerStarted","Data":"08cf6551f91428b7c627c96d8e8a3439f32ac915f3a65d321d557ccc33cbc16d"} Oct 08 14:23:38 crc kubenswrapper[4789]: I1008 14:23:38.559333 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 14:23:38 crc kubenswrapper[4789]: I1008 14:23:38.559850 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" containerID="cri-o://bbaf9b2938fb9a5809414790ba8a305f1118bf8d38764891a274c4d9bd99e4af" gracePeriod=30 Oct 08 14:23:40 crc kubenswrapper[4789]: I1008 14:23:40.239123 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:23:40 crc kubenswrapper[4789]: I1008 14:23:40.239750 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="ceilometer-central-agent" containerID="cri-o://4bef4bad303b0844f63c432c915c5fcbfaab91aa4c11130e7915d8fc8db26b61" gracePeriod=30 Oct 08 14:23:40 crc kubenswrapper[4789]: I1008 14:23:40.240285 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="proxy-httpd" containerID="cri-o://a2909f922794801e50a3e6cbd2fd155131e523d6e101ea710c21f5f5af8d2e28" gracePeriod=30 Oct 08 14:23:40 crc kubenswrapper[4789]: I1008 14:23:40.240336 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="sg-core" containerID="cri-o://2e2c20f13cc3e8474e669c53397f2098e91c7a8e073298deada1d92d32bf3b46" gracePeriod=30 Oct 08 14:23:40 crc kubenswrapper[4789]: I1008 14:23:40.240373 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="ceilometer-notification-agent" containerID="cri-o://9e91c3c4bf79c57d9547837c1ba31c3f0074f731b9965477c64868c2b0a6e30f" gracePeriod=30 Oct 08 14:23:41 crc kubenswrapper[4789]: I1008 14:23:41.067194 4789 generic.go:334] "Generic (PLEG): container finished" podID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerID="a2909f922794801e50a3e6cbd2fd155131e523d6e101ea710c21f5f5af8d2e28" exitCode=0 Oct 08 14:23:41 crc kubenswrapper[4789]: I1008 14:23:41.067228 4789 generic.go:334] "Generic (PLEG): container finished" podID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerID="2e2c20f13cc3e8474e669c53397f2098e91c7a8e073298deada1d92d32bf3b46" exitCode=2 Oct 08 14:23:41 crc kubenswrapper[4789]: I1008 14:23:41.067252 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerDied","Data":"a2909f922794801e50a3e6cbd2fd155131e523d6e101ea710c21f5f5af8d2e28"} Oct 08 14:23:41 crc kubenswrapper[4789]: I1008 14:23:41.067284 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerDied","Data":"2e2c20f13cc3e8474e669c53397f2098e91c7a8e073298deada1d92d32bf3b46"} Oct 08 14:23:42 crc kubenswrapper[4789]: I1008 14:23:42.082548 4789 generic.go:334] "Generic (PLEG): container finished" podID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerID="bbaf9b2938fb9a5809414790ba8a305f1118bf8d38764891a274c4d9bd99e4af" exitCode=0 Oct 08 14:23:42 crc kubenswrapper[4789]: I1008 14:23:42.083042 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"33bbfc8e-64bd-480c-b755-dd802cba53fe","Type":"ContainerDied","Data":"bbaf9b2938fb9a5809414790ba8a305f1118bf8d38764891a274c4d9bd99e4af"} Oct 08 14:23:42 crc kubenswrapper[4789]: I1008 14:23:42.083093 4789 scope.go:117] "RemoveContainer" containerID="440e900070209cbbb2fed3c3255ae9692fdbb53508c99ab21fd0f7653dc31a03" Oct 08 14:23:42 crc kubenswrapper[4789]: I1008 14:23:42.086501 4789 generic.go:334] "Generic (PLEG): container finished" podID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerID="4bef4bad303b0844f63c432c915c5fcbfaab91aa4c11130e7915d8fc8db26b61" exitCode=0 Oct 08 14:23:42 crc kubenswrapper[4789]: I1008 14:23:42.086540 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerDied","Data":"4bef4bad303b0844f63c432c915c5fcbfaab91aa4c11130e7915d8fc8db26b61"} Oct 08 14:23:45 crc kubenswrapper[4789]: I1008 14:23:45.474562 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:45 crc kubenswrapper[4789]: I1008 14:23:45.647486 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="a72d40d2-86d8-406f-9f5c-adec376cc947" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.195:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:23:46 crc kubenswrapper[4789]: I1008 14:23:46.645168 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="a72d40d2-86d8-406f-9f5c-adec376cc947" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.195:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.802180 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.930185 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-config-data\") pod \"33bbfc8e-64bd-480c-b755-dd802cba53fe\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.930224 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-custom-prometheus-ca\") pod \"33bbfc8e-64bd-480c-b755-dd802cba53fe\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.930266 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzcpl\" (UniqueName: \"kubernetes.io/projected/33bbfc8e-64bd-480c-b755-dd802cba53fe-kube-api-access-vzcpl\") pod \"33bbfc8e-64bd-480c-b755-dd802cba53fe\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.930358 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33bbfc8e-64bd-480c-b755-dd802cba53fe-logs\") pod \"33bbfc8e-64bd-480c-b755-dd802cba53fe\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.930400 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-combined-ca-bundle\") pod \"33bbfc8e-64bd-480c-b755-dd802cba53fe\" (UID: \"33bbfc8e-64bd-480c-b755-dd802cba53fe\") " Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.931588 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33bbfc8e-64bd-480c-b755-dd802cba53fe-logs" (OuterVolumeSpecName: "logs") pod "33bbfc8e-64bd-480c-b755-dd802cba53fe" (UID: "33bbfc8e-64bd-480c-b755-dd802cba53fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.952037 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33bbfc8e-64bd-480c-b755-dd802cba53fe-kube-api-access-vzcpl" (OuterVolumeSpecName: "kube-api-access-vzcpl") pod "33bbfc8e-64bd-480c-b755-dd802cba53fe" (UID: "33bbfc8e-64bd-480c-b755-dd802cba53fe"). InnerVolumeSpecName "kube-api-access-vzcpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.963702 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33bbfc8e-64bd-480c-b755-dd802cba53fe" (UID: "33bbfc8e-64bd-480c-b755-dd802cba53fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.970731 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "33bbfc8e-64bd-480c-b755-dd802cba53fe" (UID: "33bbfc8e-64bd-480c-b755-dd802cba53fe"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:48 crc kubenswrapper[4789]: I1008 14:23:48.992387 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-config-data" (OuterVolumeSpecName: "config-data") pod "33bbfc8e-64bd-480c-b755-dd802cba53fe" (UID: "33bbfc8e-64bd-480c-b755-dd802cba53fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.033173 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.033209 4789 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.033221 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzcpl\" (UniqueName: \"kubernetes.io/projected/33bbfc8e-64bd-480c-b755-dd802cba53fe-kube-api-access-vzcpl\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.033229 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33bbfc8e-64bd-480c-b755-dd802cba53fe-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.033238 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33bbfc8e-64bd-480c-b755-dd802cba53fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.149931 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"33bbfc8e-64bd-480c-b755-dd802cba53fe","Type":"ContainerDied","Data":"5faf3fe89fcfa6973f4667c648b25e7d874fdbd51b339cda678b294c19ca4506"} Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.149972 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.150011 4789 scope.go:117] "RemoveContainer" containerID="bbaf9b2938fb9a5809414790ba8a305f1118bf8d38764891a274c4d9bd99e4af" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.184835 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.193316 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.208898 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 14:23:49 crc kubenswrapper[4789]: E1008 14:23:49.209472 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.209489 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" Oct 08 14:23:49 crc kubenswrapper[4789]: E1008 14:23:49.209511 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.209517 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" Oct 08 14:23:49 crc kubenswrapper[4789]: E1008 14:23:49.209531 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.209537 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.209739 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.209754 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.209766 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" containerName="watcher-decision-engine" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.210498 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.213349 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.221423 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.338928 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c9e2e8c-9df6-4340-af61-2778209a54dd-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.339278 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c9e2e8c-9df6-4340-af61-2778209a54dd-logs\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.339385 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8qsr\" (UniqueName: \"kubernetes.io/projected/4c9e2e8c-9df6-4340-af61-2778209a54dd-kube-api-access-t8qsr\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.339469 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4c9e2e8c-9df6-4340-af61-2778209a54dd-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.339552 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c9e2e8c-9df6-4340-af61-2778209a54dd-config-data\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.441945 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8qsr\" (UniqueName: \"kubernetes.io/projected/4c9e2e8c-9df6-4340-af61-2778209a54dd-kube-api-access-t8qsr\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.442283 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4c9e2e8c-9df6-4340-af61-2778209a54dd-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.442399 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c9e2e8c-9df6-4340-af61-2778209a54dd-config-data\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.442608 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c9e2e8c-9df6-4340-af61-2778209a54dd-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.442740 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c9e2e8c-9df6-4340-af61-2778209a54dd-logs\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.443231 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c9e2e8c-9df6-4340-af61-2778209a54dd-logs\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.446040 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/4c9e2e8c-9df6-4340-af61-2778209a54dd-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.446834 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c9e2e8c-9df6-4340-af61-2778209a54dd-config-data\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.447157 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c9e2e8c-9df6-4340-af61-2778209a54dd-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: E1008 14:23:49.455009 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-nova-conductor:watcher_latest" Oct 08 14:23:49 crc kubenswrapper[4789]: E1008 14:23:49.455065 4789 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.222:5001/podified-master-centos10/openstack-nova-conductor:watcher_latest" Oct 08 14:23:49 crc kubenswrapper[4789]: E1008 14:23:49.455191 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:38.102.83.222:5001/podified-master-centos10/openstack-nova-conductor:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gs77f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-75hcm_openstack(c4779c61-4a62-4853-b877-fbf588daa8e8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 14:23:49 crc kubenswrapper[4789]: E1008 14:23:49.456587 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-75hcm" podUID="c4779c61-4a62-4853-b877-fbf588daa8e8" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.459191 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8qsr\" (UniqueName: \"kubernetes.io/projected/4c9e2e8c-9df6-4340-af61-2778209a54dd-kube-api-access-t8qsr\") pod \"watcher-decision-engine-0\" (UID: \"4c9e2e8c-9df6-4340-af61-2778209a54dd\") " pod="openstack/watcher-decision-engine-0" Oct 08 14:23:49 crc kubenswrapper[4789]: I1008 14:23:49.536316 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 08 14:23:50 crc kubenswrapper[4789]: E1008 14:23:50.168419 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.222:5001/podified-master-centos10/openstack-nova-conductor:watcher_latest\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-75hcm" podUID="c4779c61-4a62-4853-b877-fbf588daa8e8" Oct 08 14:23:50 crc kubenswrapper[4789]: I1008 14:23:50.216514 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 08 14:23:50 crc kubenswrapper[4789]: I1008 14:23:50.652145 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="a72d40d2-86d8-406f-9f5c-adec376cc947" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.195:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 14:23:50 crc kubenswrapper[4789]: I1008 14:23:50.753785 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33bbfc8e-64bd-480c-b755-dd802cba53fe" path="/var/lib/kubelet/pods/33bbfc8e-64bd-480c-b755-dd802cba53fe/volumes" Oct 08 14:23:51 crc kubenswrapper[4789]: I1008 14:23:51.174611 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"4c9e2e8c-9df6-4340-af61-2778209a54dd","Type":"ContainerStarted","Data":"cdf2ff539f14b346046a0f6208b35996624fae5720bdc6b66e9f914a559b2f44"} Oct 08 14:23:51 crc kubenswrapper[4789]: I1008 14:23:51.174660 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"4c9e2e8c-9df6-4340-af61-2778209a54dd","Type":"ContainerStarted","Data":"05403a01cf644e0ef2ae90406840adb012dec55c64e04fff1c24fe3467998da4"} Oct 08 14:23:51 crc kubenswrapper[4789]: I1008 14:23:51.196188 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.196170889 podStartE2EDuration="2.196170889s" podCreationTimestamp="2025-10-08 14:23:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:23:51.189712554 +0000 UTC m=+1371.096460046" watchObservedRunningTime="2025-10-08 14:23:51.196170889 +0000 UTC m=+1371.102918371" Oct 08 14:23:51 crc kubenswrapper[4789]: I1008 14:23:51.452215 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7dcb5d4fcc-q9spz" Oct 08 14:23:51 crc kubenswrapper[4789]: I1008 14:23:51.511660 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8687c9c966-lnx5h"] Oct 08 14:23:51 crc kubenswrapper[4789]: I1008 14:23:51.511920 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8687c9c966-lnx5h" podUID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerName="neutron-api" containerID="cri-o://1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f" gracePeriod=30 Oct 08 14:23:51 crc kubenswrapper[4789]: I1008 14:23:51.512317 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8687c9c966-lnx5h" podUID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerName="neutron-httpd" containerID="cri-o://462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2" gracePeriod=30 Oct 08 14:23:51 crc kubenswrapper[4789]: I1008 14:23:51.651213 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="a72d40d2-86d8-406f-9f5c-adec376cc947" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.195:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:23:53 crc kubenswrapper[4789]: I1008 14:23:53.216913 4789 generic.go:334] "Generic (PLEG): container finished" podID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerID="462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2" exitCode=0 Oct 08 14:23:53 crc kubenswrapper[4789]: I1008 14:23:53.217153 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8687c9c966-lnx5h" event={"ID":"1e5167e7-8bb8-4580-b667-d0f5083b156b","Type":"ContainerDied","Data":"462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2"} Oct 08 14:23:53 crc kubenswrapper[4789]: I1008 14:23:53.975286 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.142353 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j67tm\" (UniqueName: \"kubernetes.io/projected/1e5167e7-8bb8-4580-b667-d0f5083b156b-kube-api-access-j67tm\") pod \"1e5167e7-8bb8-4580-b667-d0f5083b156b\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.142500 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-httpd-config\") pod \"1e5167e7-8bb8-4580-b667-d0f5083b156b\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.142536 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-config\") pod \"1e5167e7-8bb8-4580-b667-d0f5083b156b\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.142604 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-ovndb-tls-certs\") pod \"1e5167e7-8bb8-4580-b667-d0f5083b156b\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.142704 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-combined-ca-bundle\") pod \"1e5167e7-8bb8-4580-b667-d0f5083b156b\" (UID: \"1e5167e7-8bb8-4580-b667-d0f5083b156b\") " Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.162459 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e5167e7-8bb8-4580-b667-d0f5083b156b-kube-api-access-j67tm" (OuterVolumeSpecName: "kube-api-access-j67tm") pod "1e5167e7-8bb8-4580-b667-d0f5083b156b" (UID: "1e5167e7-8bb8-4580-b667-d0f5083b156b"). InnerVolumeSpecName "kube-api-access-j67tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.163134 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "1e5167e7-8bb8-4580-b667-d0f5083b156b" (UID: "1e5167e7-8bb8-4580-b667-d0f5083b156b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.205141 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-config" (OuterVolumeSpecName: "config") pod "1e5167e7-8bb8-4580-b667-d0f5083b156b" (UID: "1e5167e7-8bb8-4580-b667-d0f5083b156b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.211173 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e5167e7-8bb8-4580-b667-d0f5083b156b" (UID: "1e5167e7-8bb8-4580-b667-d0f5083b156b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.226946 4789 generic.go:334] "Generic (PLEG): container finished" podID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerID="1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f" exitCode=0 Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.226998 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8687c9c966-lnx5h" event={"ID":"1e5167e7-8bb8-4580-b667-d0f5083b156b","Type":"ContainerDied","Data":"1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f"} Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.227022 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8687c9c966-lnx5h" event={"ID":"1e5167e7-8bb8-4580-b667-d0f5083b156b","Type":"ContainerDied","Data":"c9ca7edb473dd7d8142ab7390818dcd92093ad5a6c52ba2527a4b676b043f04f"} Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.227039 4789 scope.go:117] "RemoveContainer" containerID="462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.227142 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8687c9c966-lnx5h" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.229744 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "1e5167e7-8bb8-4580-b667-d0f5083b156b" (UID: "1e5167e7-8bb8-4580-b667-d0f5083b156b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.237057 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.244634 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.244664 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j67tm\" (UniqueName: \"kubernetes.io/projected/1e5167e7-8bb8-4580-b667-d0f5083b156b-kube-api-access-j67tm\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.244674 4789 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.244686 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.244698 4789 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e5167e7-8bb8-4580-b667-d0f5083b156b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.287449 4789 scope.go:117] "RemoveContainer" containerID="1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.315253 4789 scope.go:117] "RemoveContainer" containerID="462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2" Oct 08 14:23:54 crc kubenswrapper[4789]: E1008 14:23:54.315767 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2\": container with ID starting with 462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2 not found: ID does not exist" containerID="462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.315821 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2"} err="failed to get container status \"462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2\": rpc error: code = NotFound desc = could not find container \"462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2\": container with ID starting with 462c9c59fa6213167005cee7643d6c3b0ca262a2a4dc8979a6492c396835f1e2 not found: ID does not exist" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.315857 4789 scope.go:117] "RemoveContainer" containerID="1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f" Oct 08 14:23:54 crc kubenswrapper[4789]: E1008 14:23:54.316295 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f\": container with ID starting with 1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f not found: ID does not exist" containerID="1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.316323 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f"} err="failed to get container status \"1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f\": rpc error: code = NotFound desc = could not find container \"1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f\": container with ID starting with 1ed14e580998dbfa22bd3e1117193e8aefa53521b2e816aef264243a9c47b28f not found: ID does not exist" Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.558952 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8687c9c966-lnx5h"] Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.566811 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8687c9c966-lnx5h"] Oct 08 14:23:54 crc kubenswrapper[4789]: I1008 14:23:54.741585 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e5167e7-8bb8-4580-b667-d0f5083b156b" path="/var/lib/kubelet/pods/1e5167e7-8bb8-4580-b667-d0f5083b156b/volumes" Oct 08 14:23:56 crc kubenswrapper[4789]: I1008 14:23:56.432524 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:23:56 crc kubenswrapper[4789]: I1008 14:23:56.432799 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:23:56 crc kubenswrapper[4789]: I1008 14:23:56.432840 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:23:56 crc kubenswrapper[4789]: I1008 14:23:56.433561 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"26322f12c01ca82f351a4ae80e71651fb887b0762ae10b5b893ab0870c97be77"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:23:56 crc kubenswrapper[4789]: I1008 14:23:56.433614 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://26322f12c01ca82f351a4ae80e71651fb887b0762ae10b5b893ab0870c97be77" gracePeriod=600 Oct 08 14:23:56 crc kubenswrapper[4789]: E1008 14:23:56.643704 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b2e313a_834b_47fb_adae_992853bf6a6a.slice/crio-conmon-26322f12c01ca82f351a4ae80e71651fb887b0762ae10b5b893ab0870c97be77.scope\": RecentStats: unable to find data in memory cache]" Oct 08 14:23:57 crc kubenswrapper[4789]: I1008 14:23:57.224728 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.194:3000/\": dial tcp 10.217.0.194:3000: connect: connection refused" Oct 08 14:23:57 crc kubenswrapper[4789]: I1008 14:23:57.258907 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="26322f12c01ca82f351a4ae80e71651fb887b0762ae10b5b893ab0870c97be77" exitCode=0 Oct 08 14:23:57 crc kubenswrapper[4789]: I1008 14:23:57.258953 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"26322f12c01ca82f351a4ae80e71651fb887b0762ae10b5b893ab0870c97be77"} Oct 08 14:23:57 crc kubenswrapper[4789]: I1008 14:23:57.258980 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30"} Oct 08 14:23:57 crc kubenswrapper[4789]: I1008 14:23:57.259013 4789 scope.go:117] "RemoveContainer" containerID="b8550892ad865dfdb5faa70bb82d5784fbdac5d69789616846b92c98439e8b06" Oct 08 14:23:59 crc kubenswrapper[4789]: I1008 14:23:59.537046 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 08 14:23:59 crc kubenswrapper[4789]: I1008 14:23:59.571497 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 08 14:24:00 crc kubenswrapper[4789]: I1008 14:24:00.294857 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 08 14:24:00 crc kubenswrapper[4789]: I1008 14:24:00.319714 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 08 14:24:05 crc kubenswrapper[4789]: I1008 14:24:05.347009 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-75hcm" event={"ID":"c4779c61-4a62-4853-b877-fbf588daa8e8","Type":"ContainerStarted","Data":"2f006ed9c2457e36fab03b66b0e890e963638ae6dcc79112e3884b1c14a9d6d7"} Oct 08 14:24:05 crc kubenswrapper[4789]: I1008 14:24:05.371550 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-75hcm" podStartSLOduration=1.469023856 podStartE2EDuration="30.371519262s" podCreationTimestamp="2025-10-08 14:23:35 +0000 UTC" firstStartedPulling="2025-10-08 14:23:36.143511056 +0000 UTC m=+1356.050258548" lastFinishedPulling="2025-10-08 14:24:05.046006462 +0000 UTC m=+1384.952753954" observedRunningTime="2025-10-08 14:24:05.365731125 +0000 UTC m=+1385.272478617" watchObservedRunningTime="2025-10-08 14:24:05.371519262 +0000 UTC m=+1385.278266754" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.394985 4789 generic.go:334] "Generic (PLEG): container finished" podID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerID="9e91c3c4bf79c57d9547837c1ba31c3f0074f731b9965477c64868c2b0a6e30f" exitCode=137 Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.395031 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerDied","Data":"9e91c3c4bf79c57d9547837c1ba31c3f0074f731b9965477c64868c2b0a6e30f"} Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.658181 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.753591 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-run-httpd\") pod \"fa4caa03-96c7-4142-a059-8f331cbbf822\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.753681 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w4dg\" (UniqueName: \"kubernetes.io/projected/fa4caa03-96c7-4142-a059-8f331cbbf822-kube-api-access-8w4dg\") pod \"fa4caa03-96c7-4142-a059-8f331cbbf822\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.753709 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-sg-core-conf-yaml\") pod \"fa4caa03-96c7-4142-a059-8f331cbbf822\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.753764 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-config-data\") pod \"fa4caa03-96c7-4142-a059-8f331cbbf822\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.753778 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-scripts\") pod \"fa4caa03-96c7-4142-a059-8f331cbbf822\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.753843 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-combined-ca-bundle\") pod \"fa4caa03-96c7-4142-a059-8f331cbbf822\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.753878 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-log-httpd\") pod \"fa4caa03-96c7-4142-a059-8f331cbbf822\" (UID: \"fa4caa03-96c7-4142-a059-8f331cbbf822\") " Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.754310 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fa4caa03-96c7-4142-a059-8f331cbbf822" (UID: "fa4caa03-96c7-4142-a059-8f331cbbf822"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.754734 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fa4caa03-96c7-4142-a059-8f331cbbf822" (UID: "fa4caa03-96c7-4142-a059-8f331cbbf822"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.759293 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-scripts" (OuterVolumeSpecName: "scripts") pod "fa4caa03-96c7-4142-a059-8f331cbbf822" (UID: "fa4caa03-96c7-4142-a059-8f331cbbf822"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.759893 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa4caa03-96c7-4142-a059-8f331cbbf822-kube-api-access-8w4dg" (OuterVolumeSpecName: "kube-api-access-8w4dg") pod "fa4caa03-96c7-4142-a059-8f331cbbf822" (UID: "fa4caa03-96c7-4142-a059-8f331cbbf822"). InnerVolumeSpecName "kube-api-access-8w4dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.787089 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fa4caa03-96c7-4142-a059-8f331cbbf822" (UID: "fa4caa03-96c7-4142-a059-8f331cbbf822"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.831217 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa4caa03-96c7-4142-a059-8f331cbbf822" (UID: "fa4caa03-96c7-4142-a059-8f331cbbf822"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.852849 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-config-data" (OuterVolumeSpecName: "config-data") pod "fa4caa03-96c7-4142-a059-8f331cbbf822" (UID: "fa4caa03-96c7-4142-a059-8f331cbbf822"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.856443 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.856481 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.856494 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa4caa03-96c7-4142-a059-8f331cbbf822-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.856505 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w4dg\" (UniqueName: \"kubernetes.io/projected/fa4caa03-96c7-4142-a059-8f331cbbf822-kube-api-access-8w4dg\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.856519 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.856530 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:10 crc kubenswrapper[4789]: I1008 14:24:10.856540 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa4caa03-96c7-4142-a059-8f331cbbf822-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.413765 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa4caa03-96c7-4142-a059-8f331cbbf822","Type":"ContainerDied","Data":"11cb66d15cd10ee7388bfa73962fc056ca738aa7b6073194b33ea9d700711052"} Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.413827 4789 scope.go:117] "RemoveContainer" containerID="a2909f922794801e50a3e6cbd2fd155131e523d6e101ea710c21f5f5af8d2e28" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.414020 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.459375 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.473723 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.482775 4789 scope.go:117] "RemoveContainer" containerID="2e2c20f13cc3e8474e669c53397f2098e91c7a8e073298deada1d92d32bf3b46" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.492319 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:24:11 crc kubenswrapper[4789]: E1008 14:24:11.492926 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="ceilometer-notification-agent" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.492956 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="ceilometer-notification-agent" Oct 08 14:24:11 crc kubenswrapper[4789]: E1008 14:24:11.493017 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="sg-core" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.493027 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="sg-core" Oct 08 14:24:11 crc kubenswrapper[4789]: E1008 14:24:11.493038 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerName="neutron-api" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.493047 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerName="neutron-api" Oct 08 14:24:11 crc kubenswrapper[4789]: E1008 14:24:11.493057 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="proxy-httpd" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.493064 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="proxy-httpd" Oct 08 14:24:11 crc kubenswrapper[4789]: E1008 14:24:11.493090 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerName="neutron-httpd" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.493097 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerName="neutron-httpd" Oct 08 14:24:11 crc kubenswrapper[4789]: E1008 14:24:11.493112 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="ceilometer-central-agent" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.493122 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="ceilometer-central-agent" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.513118 4789 scope.go:117] "RemoveContainer" containerID="9e91c3c4bf79c57d9547837c1ba31c3f0074f731b9965477c64868c2b0a6e30f" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.513187 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerName="neutron-api" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.513585 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="proxy-httpd" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.513654 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e5167e7-8bb8-4580-b667-d0f5083b156b" containerName="neutron-httpd" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.513726 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="ceilometer-notification-agent" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.513750 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="ceilometer-central-agent" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.513770 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" containerName="sg-core" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.518785 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.525370 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.525400 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.540267 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.551866 4789 scope.go:117] "RemoveContainer" containerID="4bef4bad303b0844f63c432c915c5fcbfaab91aa4c11130e7915d8fc8db26b61" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.675521 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvcdk\" (UniqueName: \"kubernetes.io/projected/0d62e800-8147-48c2-bab8-6e16c6e22d05-kube-api-access-dvcdk\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.675799 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.675944 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-run-httpd\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.676077 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-scripts\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.676223 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-log-httpd\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.676371 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-config-data\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.676506 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.778336 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-log-httpd\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.778652 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-config-data\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.778763 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.778872 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvcdk\" (UniqueName: \"kubernetes.io/projected/0d62e800-8147-48c2-bab8-6e16c6e22d05-kube-api-access-dvcdk\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.778964 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.779076 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-run-httpd\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.778808 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-log-httpd\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.779407 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-run-httpd\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.779156 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-scripts\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.786982 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.787150 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-config-data\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.788109 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.791980 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-scripts\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.797505 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvcdk\" (UniqueName: \"kubernetes.io/projected/0d62e800-8147-48c2-bab8-6e16c6e22d05-kube-api-access-dvcdk\") pod \"ceilometer-0\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " pod="openstack/ceilometer-0" Oct 08 14:24:11 crc kubenswrapper[4789]: I1008 14:24:11.850111 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:24:12 crc kubenswrapper[4789]: I1008 14:24:12.297306 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:24:12 crc kubenswrapper[4789]: I1008 14:24:12.425387 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerStarted","Data":"82b338c9aa2efcedbe8b04682635b2523691213a211961a37c526c9bad3f239b"} Oct 08 14:24:12 crc kubenswrapper[4789]: I1008 14:24:12.744530 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa4caa03-96c7-4142-a059-8f331cbbf822" path="/var/lib/kubelet/pods/fa4caa03-96c7-4142-a059-8f331cbbf822/volumes" Oct 08 14:24:13 crc kubenswrapper[4789]: I1008 14:24:13.442705 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerStarted","Data":"1d0969b22c0457046963c721f2fe13930d401fb05ce2543511f2886ebca4be00"} Oct 08 14:24:13 crc kubenswrapper[4789]: I1008 14:24:13.443471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerStarted","Data":"21ec477643d66c1646db0374733fea8d174901474abbc958fff28425b0b3d3ee"} Oct 08 14:24:14 crc kubenswrapper[4789]: I1008 14:24:14.456025 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerStarted","Data":"0d52a25246184153286f994395ebb0b5b8722d0f40c5743dc54a7c238d1e5f3b"} Oct 08 14:24:17 crc kubenswrapper[4789]: I1008 14:24:17.482508 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerStarted","Data":"461690d543781b3c72717e5cc6694c5a7c839c5e48b71e11ee21c99d3e03b30a"} Oct 08 14:24:17 crc kubenswrapper[4789]: I1008 14:24:17.483068 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.402717 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.101304055 podStartE2EDuration="7.402700069s" podCreationTimestamp="2025-10-08 14:24:11 +0000 UTC" firstStartedPulling="2025-10-08 14:24:12.305459102 +0000 UTC m=+1392.212206594" lastFinishedPulling="2025-10-08 14:24:16.606855116 +0000 UTC m=+1396.513602608" observedRunningTime="2025-10-08 14:24:17.509268939 +0000 UTC m=+1397.416016431" watchObservedRunningTime="2025-10-08 14:24:18.402700069 +0000 UTC m=+1398.309447561" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.404248 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rg47r"] Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.406197 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.419586 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rg47r"] Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.508216 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-utilities\") pod \"redhat-marketplace-rg47r\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.508374 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-catalog-content\") pod \"redhat-marketplace-rg47r\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.508438 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9wtz\" (UniqueName: \"kubernetes.io/projected/dbc74db2-88b7-4fd9-b08d-3f6134dff432-kube-api-access-b9wtz\") pod \"redhat-marketplace-rg47r\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.610297 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-catalog-content\") pod \"redhat-marketplace-rg47r\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.610381 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9wtz\" (UniqueName: \"kubernetes.io/projected/dbc74db2-88b7-4fd9-b08d-3f6134dff432-kube-api-access-b9wtz\") pod \"redhat-marketplace-rg47r\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.610467 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-utilities\") pod \"redhat-marketplace-rg47r\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.611633 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-catalog-content\") pod \"redhat-marketplace-rg47r\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.611868 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-utilities\") pod \"redhat-marketplace-rg47r\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.631627 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9wtz\" (UniqueName: \"kubernetes.io/projected/dbc74db2-88b7-4fd9-b08d-3f6134dff432-kube-api-access-b9wtz\") pod \"redhat-marketplace-rg47r\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:18 crc kubenswrapper[4789]: I1008 14:24:18.726084 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:19 crc kubenswrapper[4789]: I1008 14:24:19.209136 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rg47r"] Oct 08 14:24:19 crc kubenswrapper[4789]: W1008 14:24:19.218055 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbc74db2_88b7_4fd9_b08d_3f6134dff432.slice/crio-f6433a920df86770695f18aae52797780d7f27d0a838429b01b0442f0b50651c WatchSource:0}: Error finding container f6433a920df86770695f18aae52797780d7f27d0a838429b01b0442f0b50651c: Status 404 returned error can't find the container with id f6433a920df86770695f18aae52797780d7f27d0a838429b01b0442f0b50651c Oct 08 14:24:19 crc kubenswrapper[4789]: I1008 14:24:19.502640 4789 generic.go:334] "Generic (PLEG): container finished" podID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerID="abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a" exitCode=0 Oct 08 14:24:19 crc kubenswrapper[4789]: I1008 14:24:19.502755 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rg47r" event={"ID":"dbc74db2-88b7-4fd9-b08d-3f6134dff432","Type":"ContainerDied","Data":"abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a"} Oct 08 14:24:19 crc kubenswrapper[4789]: I1008 14:24:19.503012 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rg47r" event={"ID":"dbc74db2-88b7-4fd9-b08d-3f6134dff432","Type":"ContainerStarted","Data":"f6433a920df86770695f18aae52797780d7f27d0a838429b01b0442f0b50651c"} Oct 08 14:24:21 crc kubenswrapper[4789]: I1008 14:24:21.521073 4789 generic.go:334] "Generic (PLEG): container finished" podID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerID="b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1" exitCode=0 Oct 08 14:24:21 crc kubenswrapper[4789]: I1008 14:24:21.521498 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rg47r" event={"ID":"dbc74db2-88b7-4fd9-b08d-3f6134dff432","Type":"ContainerDied","Data":"b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1"} Oct 08 14:24:25 crc kubenswrapper[4789]: I1008 14:24:25.558195 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rg47r" event={"ID":"dbc74db2-88b7-4fd9-b08d-3f6134dff432","Type":"ContainerStarted","Data":"157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a"} Oct 08 14:24:25 crc kubenswrapper[4789]: I1008 14:24:25.559613 4789 generic.go:334] "Generic (PLEG): container finished" podID="c4779c61-4a62-4853-b877-fbf588daa8e8" containerID="2f006ed9c2457e36fab03b66b0e890e963638ae6dcc79112e3884b1c14a9d6d7" exitCode=0 Oct 08 14:24:25 crc kubenswrapper[4789]: I1008 14:24:25.560102 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-75hcm" event={"ID":"c4779c61-4a62-4853-b877-fbf588daa8e8","Type":"ContainerDied","Data":"2f006ed9c2457e36fab03b66b0e890e963638ae6dcc79112e3884b1c14a9d6d7"} Oct 08 14:24:25 crc kubenswrapper[4789]: I1008 14:24:25.578625 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rg47r" podStartSLOduration=2.627270208 podStartE2EDuration="7.578598934s" podCreationTimestamp="2025-10-08 14:24:18 +0000 UTC" firstStartedPulling="2025-10-08 14:24:19.504302369 +0000 UTC m=+1399.411049861" lastFinishedPulling="2025-10-08 14:24:24.455631095 +0000 UTC m=+1404.362378587" observedRunningTime="2025-10-08 14:24:25.576253081 +0000 UTC m=+1405.483000573" watchObservedRunningTime="2025-10-08 14:24:25.578598934 +0000 UTC m=+1405.485346446" Oct 08 14:24:26 crc kubenswrapper[4789]: I1008 14:24:26.963287 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.079851 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-combined-ca-bundle\") pod \"c4779c61-4a62-4853-b877-fbf588daa8e8\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.079911 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs77f\" (UniqueName: \"kubernetes.io/projected/c4779c61-4a62-4853-b877-fbf588daa8e8-kube-api-access-gs77f\") pod \"c4779c61-4a62-4853-b877-fbf588daa8e8\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.080107 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-config-data\") pod \"c4779c61-4a62-4853-b877-fbf588daa8e8\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.080174 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-scripts\") pod \"c4779c61-4a62-4853-b877-fbf588daa8e8\" (UID: \"c4779c61-4a62-4853-b877-fbf588daa8e8\") " Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.089696 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4779c61-4a62-4853-b877-fbf588daa8e8-kube-api-access-gs77f" (OuterVolumeSpecName: "kube-api-access-gs77f") pod "c4779c61-4a62-4853-b877-fbf588daa8e8" (UID: "c4779c61-4a62-4853-b877-fbf588daa8e8"). InnerVolumeSpecName "kube-api-access-gs77f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.095507 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-scripts" (OuterVolumeSpecName: "scripts") pod "c4779c61-4a62-4853-b877-fbf588daa8e8" (UID: "c4779c61-4a62-4853-b877-fbf588daa8e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.112505 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4779c61-4a62-4853-b877-fbf588daa8e8" (UID: "c4779c61-4a62-4853-b877-fbf588daa8e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.112832 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-config-data" (OuterVolumeSpecName: "config-data") pod "c4779c61-4a62-4853-b877-fbf588daa8e8" (UID: "c4779c61-4a62-4853-b877-fbf588daa8e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.181711 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs77f\" (UniqueName: \"kubernetes.io/projected/c4779c61-4a62-4853-b877-fbf588daa8e8-kube-api-access-gs77f\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.181968 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.181978 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.182001 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4779c61-4a62-4853-b877-fbf588daa8e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.577914 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-75hcm" event={"ID":"c4779c61-4a62-4853-b877-fbf588daa8e8","Type":"ContainerDied","Data":"08cf6551f91428b7c627c96d8e8a3439f32ac915f3a65d321d557ccc33cbc16d"} Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.577964 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08cf6551f91428b7c627c96d8e8a3439f32ac915f3a65d321d557ccc33cbc16d" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.578022 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-75hcm" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.692908 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 14:24:27 crc kubenswrapper[4789]: E1008 14:24:27.693444 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4779c61-4a62-4853-b877-fbf588daa8e8" containerName="nova-cell0-conductor-db-sync" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.693468 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4779c61-4a62-4853-b877-fbf588daa8e8" containerName="nova-cell0-conductor-db-sync" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.693727 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4779c61-4a62-4853-b877-fbf588daa8e8" containerName="nova-cell0-conductor-db-sync" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.694536 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.697165 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wwghs" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.698772 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.702151 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.791778 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1fa794-a227-446e-a3b0-2b1c120e4ecf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"de1fa794-a227-446e-a3b0-2b1c120e4ecf\") " pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.791931 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sbmh\" (UniqueName: \"kubernetes.io/projected/de1fa794-a227-446e-a3b0-2b1c120e4ecf-kube-api-access-7sbmh\") pod \"nova-cell0-conductor-0\" (UID: \"de1fa794-a227-446e-a3b0-2b1c120e4ecf\") " pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.792356 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1fa794-a227-446e-a3b0-2b1c120e4ecf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"de1fa794-a227-446e-a3b0-2b1c120e4ecf\") " pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.894290 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1fa794-a227-446e-a3b0-2b1c120e4ecf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"de1fa794-a227-446e-a3b0-2b1c120e4ecf\") " pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.894456 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1fa794-a227-446e-a3b0-2b1c120e4ecf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"de1fa794-a227-446e-a3b0-2b1c120e4ecf\") " pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.894555 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sbmh\" (UniqueName: \"kubernetes.io/projected/de1fa794-a227-446e-a3b0-2b1c120e4ecf-kube-api-access-7sbmh\") pod \"nova-cell0-conductor-0\" (UID: \"de1fa794-a227-446e-a3b0-2b1c120e4ecf\") " pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.899195 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1fa794-a227-446e-a3b0-2b1c120e4ecf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"de1fa794-a227-446e-a3b0-2b1c120e4ecf\") " pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.909498 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1fa794-a227-446e-a3b0-2b1c120e4ecf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"de1fa794-a227-446e-a3b0-2b1c120e4ecf\") " pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:27 crc kubenswrapper[4789]: I1008 14:24:27.912654 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sbmh\" (UniqueName: \"kubernetes.io/projected/de1fa794-a227-446e-a3b0-2b1c120e4ecf-kube-api-access-7sbmh\") pod \"nova-cell0-conductor-0\" (UID: \"de1fa794-a227-446e-a3b0-2b1c120e4ecf\") " pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:28 crc kubenswrapper[4789]: I1008 14:24:28.010568 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:28 crc kubenswrapper[4789]: I1008 14:24:28.450099 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 14:24:28 crc kubenswrapper[4789]: W1008 14:24:28.452539 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde1fa794_a227_446e_a3b0_2b1c120e4ecf.slice/crio-2c52fb9b3dedded2a51a34625322e26d03891a0b0d393bffe99aea042f8834e2 WatchSource:0}: Error finding container 2c52fb9b3dedded2a51a34625322e26d03891a0b0d393bffe99aea042f8834e2: Status 404 returned error can't find the container with id 2c52fb9b3dedded2a51a34625322e26d03891a0b0d393bffe99aea042f8834e2 Oct 08 14:24:28 crc kubenswrapper[4789]: I1008 14:24:28.594956 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"de1fa794-a227-446e-a3b0-2b1c120e4ecf","Type":"ContainerStarted","Data":"2c52fb9b3dedded2a51a34625322e26d03891a0b0d393bffe99aea042f8834e2"} Oct 08 14:24:28 crc kubenswrapper[4789]: I1008 14:24:28.726674 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:28 crc kubenswrapper[4789]: I1008 14:24:28.726739 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:28 crc kubenswrapper[4789]: I1008 14:24:28.787916 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:29 crc kubenswrapper[4789]: I1008 14:24:29.605924 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"de1fa794-a227-446e-a3b0-2b1c120e4ecf","Type":"ContainerStarted","Data":"c2d93dfadfaa2009f0c05c3fb17d77d17e1e5b32b957c18f3978c9ca04bee5a2"} Oct 08 14:24:29 crc kubenswrapper[4789]: I1008 14:24:29.606314 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:29 crc kubenswrapper[4789]: I1008 14:24:29.625323 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.625303298 podStartE2EDuration="2.625303298s" podCreationTimestamp="2025-10-08 14:24:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:24:29.622088381 +0000 UTC m=+1409.528835893" watchObservedRunningTime="2025-10-08 14:24:29.625303298 +0000 UTC m=+1409.532050800" Oct 08 14:24:29 crc kubenswrapper[4789]: I1008 14:24:29.658926 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:29 crc kubenswrapper[4789]: I1008 14:24:29.710595 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rg47r"] Oct 08 14:24:31 crc kubenswrapper[4789]: I1008 14:24:31.658348 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rg47r" podUID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerName="registry-server" containerID="cri-o://157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a" gracePeriod=2 Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.156635 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.271779 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-catalog-content\") pod \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.271828 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-utilities\") pod \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.272031 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9wtz\" (UniqueName: \"kubernetes.io/projected/dbc74db2-88b7-4fd9-b08d-3f6134dff432-kube-api-access-b9wtz\") pod \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\" (UID: \"dbc74db2-88b7-4fd9-b08d-3f6134dff432\") " Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.272923 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-utilities" (OuterVolumeSpecName: "utilities") pod "dbc74db2-88b7-4fd9-b08d-3f6134dff432" (UID: "dbc74db2-88b7-4fd9-b08d-3f6134dff432"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.277832 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbc74db2-88b7-4fd9-b08d-3f6134dff432-kube-api-access-b9wtz" (OuterVolumeSpecName: "kube-api-access-b9wtz") pod "dbc74db2-88b7-4fd9-b08d-3f6134dff432" (UID: "dbc74db2-88b7-4fd9-b08d-3f6134dff432"). InnerVolumeSpecName "kube-api-access-b9wtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.284299 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbc74db2-88b7-4fd9-b08d-3f6134dff432" (UID: "dbc74db2-88b7-4fd9-b08d-3f6134dff432"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.374287 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9wtz\" (UniqueName: \"kubernetes.io/projected/dbc74db2-88b7-4fd9-b08d-3f6134dff432-kube-api-access-b9wtz\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.374331 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.374344 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbc74db2-88b7-4fd9-b08d-3f6134dff432-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.671516 4789 generic.go:334] "Generic (PLEG): container finished" podID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerID="157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a" exitCode=0 Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.671606 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rg47r" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.671617 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rg47r" event={"ID":"dbc74db2-88b7-4fd9-b08d-3f6134dff432","Type":"ContainerDied","Data":"157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a"} Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.671951 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rg47r" event={"ID":"dbc74db2-88b7-4fd9-b08d-3f6134dff432","Type":"ContainerDied","Data":"f6433a920df86770695f18aae52797780d7f27d0a838429b01b0442f0b50651c"} Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.671973 4789 scope.go:117] "RemoveContainer" containerID="157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.709307 4789 scope.go:117] "RemoveContainer" containerID="b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.709613 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rg47r"] Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.717718 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rg47r"] Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.735693 4789 scope.go:117] "RemoveContainer" containerID="abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.748572 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" path="/var/lib/kubelet/pods/dbc74db2-88b7-4fd9-b08d-3f6134dff432/volumes" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.770556 4789 scope.go:117] "RemoveContainer" containerID="157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a" Oct 08 14:24:32 crc kubenswrapper[4789]: E1008 14:24:32.771355 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a\": container with ID starting with 157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a not found: ID does not exist" containerID="157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.771474 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a"} err="failed to get container status \"157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a\": rpc error: code = NotFound desc = could not find container \"157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a\": container with ID starting with 157c6f6e26de7b39152d796a0ea4309b8d1593161c986311124a177f2a912a5a not found: ID does not exist" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.771515 4789 scope.go:117] "RemoveContainer" containerID="b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1" Oct 08 14:24:32 crc kubenswrapper[4789]: E1008 14:24:32.772058 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1\": container with ID starting with b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1 not found: ID does not exist" containerID="b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.772154 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1"} err="failed to get container status \"b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1\": rpc error: code = NotFound desc = could not find container \"b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1\": container with ID starting with b30c5206a8778742d94219dd7e2ab47fafd658c9fb595dc5daecf161e8d8a1d1 not found: ID does not exist" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.772249 4789 scope.go:117] "RemoveContainer" containerID="abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a" Oct 08 14:24:32 crc kubenswrapper[4789]: E1008 14:24:32.772593 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a\": container with ID starting with abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a not found: ID does not exist" containerID="abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a" Oct 08 14:24:32 crc kubenswrapper[4789]: I1008 14:24:32.772632 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a"} err="failed to get container status \"abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a\": rpc error: code = NotFound desc = could not find container \"abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a\": container with ID starting with abacbacb4b8376b54706b3fe6472c57dd0107d0e2858d96410fe33767cd1c77a not found: ID does not exist" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.037959 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.463212 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-v56bp"] Oct 08 14:24:33 crc kubenswrapper[4789]: E1008 14:24:33.463737 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerName="extract-utilities" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.463765 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerName="extract-utilities" Oct 08 14:24:33 crc kubenswrapper[4789]: E1008 14:24:33.463780 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerName="extract-content" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.463789 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerName="extract-content" Oct 08 14:24:33 crc kubenswrapper[4789]: E1008 14:24:33.463822 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerName="registry-server" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.463831 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerName="registry-server" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.464093 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbc74db2-88b7-4fd9-b08d-3f6134dff432" containerName="registry-server" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.465126 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.470465 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.470713 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.475672 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-v56bp"] Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.598223 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-scripts\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.598328 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b9n7\" (UniqueName: \"kubernetes.io/projected/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-kube-api-access-9b9n7\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.598376 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.598430 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-config-data\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.648109 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.650165 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.653361 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.669980 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.703500 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b9n7\" (UniqueName: \"kubernetes.io/projected/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-kube-api-access-9b9n7\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.703613 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.703724 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-config-data\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.703822 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-scripts\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.716127 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.718130 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-config-data\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.742672 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-scripts\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.742732 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.744065 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.759381 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.761843 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b9n7\" (UniqueName: \"kubernetes.io/projected/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-kube-api-access-9b9n7\") pod \"nova-cell0-cell-mapping-v56bp\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.763821 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.788564 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.839381 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88bebca2-d7f3-436d-b034-69579813cbc9-logs\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.839579 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.839619 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kh4n\" (UniqueName: \"kubernetes.io/projected/88bebca2-d7f3-436d-b034-69579813cbc9-kube-api-access-2kh4n\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.839639 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.839672 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-config-data\") pod \"nova-scheduler-0\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.839703 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf8rv\" (UniqueName: \"kubernetes.io/projected/c03e9357-b2ad-421e-96a7-d513673139a1-kube-api-access-tf8rv\") pod \"nova-scheduler-0\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.839778 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-config-data\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.933457 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.936651 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.943491 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.948972 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.949027 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kh4n\" (UniqueName: \"kubernetes.io/projected/88bebca2-d7f3-436d-b034-69579813cbc9-kube-api-access-2kh4n\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.949051 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.949084 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-config-data\") pod \"nova-scheduler-0\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.949175 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf8rv\" (UniqueName: \"kubernetes.io/projected/c03e9357-b2ad-421e-96a7-d513673139a1-kube-api-access-tf8rv\") pod \"nova-scheduler-0\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.949273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-config-data\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.949457 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88bebca2-d7f3-436d-b034-69579813cbc9-logs\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.954595 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88bebca2-d7f3-436d-b034-69579813cbc9-logs\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.972719 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.973203 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.975390 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-config-data\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.991543 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.992501 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kh4n\" (UniqueName: \"kubernetes.io/projected/88bebca2-d7f3-436d-b034-69579813cbc9-kube-api-access-2kh4n\") pod \"nova-api-0\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " pod="openstack/nova-api-0" Oct 08 14:24:33 crc kubenswrapper[4789]: I1008 14:24:33.993529 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-config-data\") pod \"nova-scheduler-0\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.007432 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf8rv\" (UniqueName: \"kubernetes.io/projected/c03e9357-b2ad-421e-96a7-d513673139a1-kube-api-access-tf8rv\") pod \"nova-scheduler-0\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.036299 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.037773 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.040188 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.059270 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.060240 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljv7m\" (UniqueName: \"kubernetes.io/projected/47f01d53-0d16-45cc-a6a7-f6ec90403969-kube-api-access-ljv7m\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.060413 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.060496 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-config-data\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.060517 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f01d53-0d16-45cc-a6a7-f6ec90403969-logs\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.079649 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65579c8c57-22dsv"] Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.081955 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.112419 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65579c8c57-22dsv"] Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.161975 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.162054 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5sdp\" (UniqueName: \"kubernetes.io/projected/33326886-bef2-4eb9-903c-79fd1a9f8779-kube-api-access-h5sdp\") pod \"nova-cell1-novncproxy-0\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.162086 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-config-data\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.162101 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f01d53-0d16-45cc-a6a7-f6ec90403969-logs\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.162116 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.162208 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.162271 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljv7m\" (UniqueName: \"kubernetes.io/projected/47f01d53-0d16-45cc-a6a7-f6ec90403969-kube-api-access-ljv7m\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.162935 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f01d53-0d16-45cc-a6a7-f6ec90403969-logs\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.168136 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.171466 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.183242 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-config-data\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.191788 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljv7m\" (UniqueName: \"kubernetes.io/projected/47f01d53-0d16-45cc-a6a7-f6ec90403969-kube-api-access-ljv7m\") pod \"nova-metadata-0\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.268059 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5sdp\" (UniqueName: \"kubernetes.io/projected/33326886-bef2-4eb9-903c-79fd1a9f8779-kube-api-access-h5sdp\") pod \"nova-cell1-novncproxy-0\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.268350 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-nb\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.268385 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.268417 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.268451 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-config\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.268474 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-svc\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.268553 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn67j\" (UniqueName: \"kubernetes.io/projected/0928b734-4a8d-46df-8726-420d02c65243-kube-api-access-sn67j\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.268572 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-sb\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.268614 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-swift-storage-0\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.270815 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.279159 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.279734 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.330633 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5sdp\" (UniqueName: \"kubernetes.io/projected/33326886-bef2-4eb9-903c-79fd1a9f8779-kube-api-access-h5sdp\") pod \"nova-cell1-novncproxy-0\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.376091 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-config\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.376150 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-svc\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.376245 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn67j\" (UniqueName: \"kubernetes.io/projected/0928b734-4a8d-46df-8726-420d02c65243-kube-api-access-sn67j\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.376263 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-sb\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.376296 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-swift-storage-0\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.376327 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-nb\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.377267 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-nb\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.377750 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-config\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.378303 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-svc\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.379105 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-sb\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.379719 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-swift-storage-0\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.408866 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn67j\" (UniqueName: \"kubernetes.io/projected/0928b734-4a8d-46df-8726-420d02c65243-kube-api-access-sn67j\") pod \"dnsmasq-dns-65579c8c57-22dsv\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.430298 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.459846 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.472541 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.525792 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-v56bp"] Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.720724 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v56bp" event={"ID":"829d0843-9f3f-4bd4-b5e3-4562c7405f5f","Type":"ContainerStarted","Data":"29d3a3c8986848772d6650b2071d0b3c9d56b9e5d9731b97e257c55359099afa"} Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.759568 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dr8h4"] Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.761090 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.769820 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.770232 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.770830 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dr8h4"] Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.899196 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-scripts\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.899261 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzb96\" (UniqueName: \"kubernetes.io/projected/4aa68379-349e-48f5-b6c5-d2b694639e9f-kube-api-access-qzb96\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.899312 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-config-data\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.899380 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:34 crc kubenswrapper[4789]: I1008 14:24:34.963957 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:24:34 crc kubenswrapper[4789]: W1008 14:24:34.985736 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc03e9357_b2ad_421e_96a7_d513673139a1.slice/crio-447961291e5f556e48f377196d497a70c0ff97ae6911fbe33ed084f5b384279e WatchSource:0}: Error finding container 447961291e5f556e48f377196d497a70c0ff97ae6911fbe33ed084f5b384279e: Status 404 returned error can't find the container with id 447961291e5f556e48f377196d497a70c0ff97ae6911fbe33ed084f5b384279e Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.002723 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.002938 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-scripts\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.003019 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzb96\" (UniqueName: \"kubernetes.io/projected/4aa68379-349e-48f5-b6c5-d2b694639e9f-kube-api-access-qzb96\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.003095 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-config-data\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.009751 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.013610 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-scripts\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.013891 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-config-data\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.021758 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzb96\" (UniqueName: \"kubernetes.io/projected/4aa68379-349e-48f5-b6c5-d2b694639e9f-kube-api-access-qzb96\") pod \"nova-cell1-conductor-db-sync-dr8h4\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.088160 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.121273 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:24:35 crc kubenswrapper[4789]: W1008 14:24:35.153477 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88bebca2_d7f3_436d_b034_69579813cbc9.slice/crio-91985a24003611da96a389a0236bdd5858c928351c8b2e4b420a2304e7baf244 WatchSource:0}: Error finding container 91985a24003611da96a389a0236bdd5858c928351c8b2e4b420a2304e7baf244: Status 404 returned error can't find the container with id 91985a24003611da96a389a0236bdd5858c928351c8b2e4b420a2304e7baf244 Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.253397 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.263940 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.377524 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65579c8c57-22dsv"] Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.612130 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dr8h4"] Oct 08 14:24:35 crc kubenswrapper[4789]: W1008 14:24:35.625225 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aa68379_349e_48f5_b6c5_d2b694639e9f.slice/crio-a3cb9561eb4726068aa7b7ea4e07b8f551e6308e2bfe71dba5989bbc71271930 WatchSource:0}: Error finding container a3cb9561eb4726068aa7b7ea4e07b8f551e6308e2bfe71dba5989bbc71271930: Status 404 returned error can't find the container with id a3cb9561eb4726068aa7b7ea4e07b8f551e6308e2bfe71dba5989bbc71271930 Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.752873 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"33326886-bef2-4eb9-903c-79fd1a9f8779","Type":"ContainerStarted","Data":"10463ecd989e1c45218ac5336869e47de6292a36c966f9215c27f819fa815ecd"} Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.761316 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v56bp" event={"ID":"829d0843-9f3f-4bd4-b5e3-4562c7405f5f","Type":"ContainerStarted","Data":"508deddf75c87868a482535af0949a762abe0b0521a1144c6bb3ee5c019f6ca3"} Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.771898 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dr8h4" event={"ID":"4aa68379-349e-48f5-b6c5-d2b694639e9f","Type":"ContainerStarted","Data":"a3cb9561eb4726068aa7b7ea4e07b8f551e6308e2bfe71dba5989bbc71271930"} Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.773639 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88bebca2-d7f3-436d-b034-69579813cbc9","Type":"ContainerStarted","Data":"91985a24003611da96a389a0236bdd5858c928351c8b2e4b420a2304e7baf244"} Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.781257 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-v56bp" podStartSLOduration=2.781239786 podStartE2EDuration="2.781239786s" podCreationTimestamp="2025-10-08 14:24:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:24:35.777067542 +0000 UTC m=+1415.683815054" watchObservedRunningTime="2025-10-08 14:24:35.781239786 +0000 UTC m=+1415.687987278" Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.785126 4789 generic.go:334] "Generic (PLEG): container finished" podID="0928b734-4a8d-46df-8726-420d02c65243" containerID="8523752a7d14be02c9e4f6b0738d78a955f5224a6e8cfa1d59c45e808998cbcd" exitCode=0 Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.785255 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" event={"ID":"0928b734-4a8d-46df-8726-420d02c65243","Type":"ContainerDied","Data":"8523752a7d14be02c9e4f6b0738d78a955f5224a6e8cfa1d59c45e808998cbcd"} Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.785289 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" event={"ID":"0928b734-4a8d-46df-8726-420d02c65243","Type":"ContainerStarted","Data":"596f0d3c7bb6bbf347dd7b30e80a116a3094d0ac055075d5bfc1db68d283d1c1"} Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.789031 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c03e9357-b2ad-421e-96a7-d513673139a1","Type":"ContainerStarted","Data":"447961291e5f556e48f377196d497a70c0ff97ae6911fbe33ed084f5b384279e"} Oct 08 14:24:35 crc kubenswrapper[4789]: I1008 14:24:35.794268 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47f01d53-0d16-45cc-a6a7-f6ec90403969","Type":"ContainerStarted","Data":"bd41e8ee5d05b140a9dc5c508af6789eeab96cfc37b0318d86ec4ccc120fb902"} Oct 08 14:24:36 crc kubenswrapper[4789]: I1008 14:24:36.811290 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" event={"ID":"0928b734-4a8d-46df-8726-420d02c65243","Type":"ContainerStarted","Data":"c752a74e4590f8230f08e6fb24649f87ad81b1253807003ac354a092dfce0b6a"} Oct 08 14:24:36 crc kubenswrapper[4789]: I1008 14:24:36.811597 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:36 crc kubenswrapper[4789]: I1008 14:24:36.815501 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dr8h4" event={"ID":"4aa68379-349e-48f5-b6c5-d2b694639e9f","Type":"ContainerStarted","Data":"ccbe308e2801588c3627cf00337a6708ad65f29cc86ef71ba89a14e63d4ee497"} Oct 08 14:24:36 crc kubenswrapper[4789]: I1008 14:24:36.832465 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" podStartSLOduration=3.832449982 podStartE2EDuration="3.832449982s" podCreationTimestamp="2025-10-08 14:24:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:24:36.829328897 +0000 UTC m=+1416.736076389" watchObservedRunningTime="2025-10-08 14:24:36.832449982 +0000 UTC m=+1416.739197474" Oct 08 14:24:36 crc kubenswrapper[4789]: I1008 14:24:36.847311 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-dr8h4" podStartSLOduration=2.847293024 podStartE2EDuration="2.847293024s" podCreationTimestamp="2025-10-08 14:24:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:24:36.843594364 +0000 UTC m=+1416.750341856" watchObservedRunningTime="2025-10-08 14:24:36.847293024 +0000 UTC m=+1416.754040516" Oct 08 14:24:37 crc kubenswrapper[4789]: I1008 14:24:37.806346 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 14:24:37 crc kubenswrapper[4789]: I1008 14:24:37.818941 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.852146 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47f01d53-0d16-45cc-a6a7-f6ec90403969","Type":"ContainerStarted","Data":"412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593"} Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.852903 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47f01d53-0d16-45cc-a6a7-f6ec90403969","Type":"ContainerStarted","Data":"324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851"} Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.852796 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerName="nova-metadata-metadata" containerID="cri-o://412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593" gracePeriod=30 Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.852226 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerName="nova-metadata-log" containerID="cri-o://324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851" gracePeriod=30 Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.854298 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"33326886-bef2-4eb9-903c-79fd1a9f8779","Type":"ContainerStarted","Data":"37f67a89188796a084b2eee0bc252d971771f456ea1e51936cc40b777fd24880"} Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.854375 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="33326886-bef2-4eb9-903c-79fd1a9f8779" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://37f67a89188796a084b2eee0bc252d971771f456ea1e51936cc40b777fd24880" gracePeriod=30 Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.858866 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88bebca2-d7f3-436d-b034-69579813cbc9","Type":"ContainerStarted","Data":"d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98"} Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.858912 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88bebca2-d7f3-436d-b034-69579813cbc9","Type":"ContainerStarted","Data":"4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311"} Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.863721 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c03e9357-b2ad-421e-96a7-d513673139a1","Type":"ContainerStarted","Data":"fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96"} Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.871107 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.279741784 podStartE2EDuration="6.871091501s" podCreationTimestamp="2025-10-08 14:24:33 +0000 UTC" firstStartedPulling="2025-10-08 14:24:35.295453432 +0000 UTC m=+1415.202200924" lastFinishedPulling="2025-10-08 14:24:38.886803149 +0000 UTC m=+1418.793550641" observedRunningTime="2025-10-08 14:24:39.869604391 +0000 UTC m=+1419.776351883" watchObservedRunningTime="2025-10-08 14:24:39.871091501 +0000 UTC m=+1419.777838993" Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.890379 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.276346922 podStartE2EDuration="6.890360813s" podCreationTimestamp="2025-10-08 14:24:33 +0000 UTC" firstStartedPulling="2025-10-08 14:24:35.274409902 +0000 UTC m=+1415.181157394" lastFinishedPulling="2025-10-08 14:24:38.888423803 +0000 UTC m=+1418.795171285" observedRunningTime="2025-10-08 14:24:39.887515416 +0000 UTC m=+1419.794262908" watchObservedRunningTime="2025-10-08 14:24:39.890360813 +0000 UTC m=+1419.797108305" Oct 08 14:24:39 crc kubenswrapper[4789]: I1008 14:24:39.917102 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.020473228 podStartE2EDuration="6.917080367s" podCreationTimestamp="2025-10-08 14:24:33 +0000 UTC" firstStartedPulling="2025-10-08 14:24:34.991801724 +0000 UTC m=+1414.898549216" lastFinishedPulling="2025-10-08 14:24:38.888408863 +0000 UTC m=+1418.795156355" observedRunningTime="2025-10-08 14:24:39.906979663 +0000 UTC m=+1419.813727155" watchObservedRunningTime="2025-10-08 14:24:39.917080367 +0000 UTC m=+1419.823827859" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.475861 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.501072 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.7361434510000002 podStartE2EDuration="7.501047651s" podCreationTimestamp="2025-10-08 14:24:33 +0000 UTC" firstStartedPulling="2025-10-08 14:24:35.156699642 +0000 UTC m=+1415.063447144" lastFinishedPulling="2025-10-08 14:24:38.921603852 +0000 UTC m=+1418.828351344" observedRunningTime="2025-10-08 14:24:39.927281464 +0000 UTC m=+1419.834028956" watchObservedRunningTime="2025-10-08 14:24:40.501047651 +0000 UTC m=+1420.407795153" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.523572 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f01d53-0d16-45cc-a6a7-f6ec90403969-logs\") pod \"47f01d53-0d16-45cc-a6a7-f6ec90403969\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.523808 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-config-data\") pod \"47f01d53-0d16-45cc-a6a7-f6ec90403969\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.523885 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-combined-ca-bundle\") pod \"47f01d53-0d16-45cc-a6a7-f6ec90403969\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.523947 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljv7m\" (UniqueName: \"kubernetes.io/projected/47f01d53-0d16-45cc-a6a7-f6ec90403969-kube-api-access-ljv7m\") pod \"47f01d53-0d16-45cc-a6a7-f6ec90403969\" (UID: \"47f01d53-0d16-45cc-a6a7-f6ec90403969\") " Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.524049 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47f01d53-0d16-45cc-a6a7-f6ec90403969-logs" (OuterVolumeSpecName: "logs") pod "47f01d53-0d16-45cc-a6a7-f6ec90403969" (UID: "47f01d53-0d16-45cc-a6a7-f6ec90403969"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.524376 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f01d53-0d16-45cc-a6a7-f6ec90403969-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.533425 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47f01d53-0d16-45cc-a6a7-f6ec90403969-kube-api-access-ljv7m" (OuterVolumeSpecName: "kube-api-access-ljv7m") pod "47f01d53-0d16-45cc-a6a7-f6ec90403969" (UID: "47f01d53-0d16-45cc-a6a7-f6ec90403969"). InnerVolumeSpecName "kube-api-access-ljv7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.565057 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-config-data" (OuterVolumeSpecName: "config-data") pod "47f01d53-0d16-45cc-a6a7-f6ec90403969" (UID: "47f01d53-0d16-45cc-a6a7-f6ec90403969"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.565342 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47f01d53-0d16-45cc-a6a7-f6ec90403969" (UID: "47f01d53-0d16-45cc-a6a7-f6ec90403969"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.625748 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.625783 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f01d53-0d16-45cc-a6a7-f6ec90403969-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.625792 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljv7m\" (UniqueName: \"kubernetes.io/projected/47f01d53-0d16-45cc-a6a7-f6ec90403969-kube-api-access-ljv7m\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.877108 4789 generic.go:334] "Generic (PLEG): container finished" podID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerID="412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593" exitCode=0 Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.877137 4789 generic.go:334] "Generic (PLEG): container finished" podID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerID="324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851" exitCode=143 Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.877150 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.877193 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47f01d53-0d16-45cc-a6a7-f6ec90403969","Type":"ContainerDied","Data":"412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593"} Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.877232 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47f01d53-0d16-45cc-a6a7-f6ec90403969","Type":"ContainerDied","Data":"324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851"} Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.877242 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47f01d53-0d16-45cc-a6a7-f6ec90403969","Type":"ContainerDied","Data":"bd41e8ee5d05b140a9dc5c508af6789eeab96cfc37b0318d86ec4ccc120fb902"} Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.877258 4789 scope.go:117] "RemoveContainer" containerID="412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.917239 4789 scope.go:117] "RemoveContainer" containerID="324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.923801 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.941400 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.952502 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:40 crc kubenswrapper[4789]: E1008 14:24:40.953907 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerName="nova-metadata-metadata" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.954001 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerName="nova-metadata-metadata" Oct 08 14:24:40 crc kubenswrapper[4789]: E1008 14:24:40.954025 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerName="nova-metadata-log" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.954033 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerName="nova-metadata-log" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.954588 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerName="nova-metadata-log" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.954654 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="47f01d53-0d16-45cc-a6a7-f6ec90403969" containerName="nova-metadata-metadata" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.956431 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.959419 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.959648 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 14:24:40 crc kubenswrapper[4789]: I1008 14:24:40.969752 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.002366 4789 scope.go:117] "RemoveContainer" containerID="412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593" Oct 08 14:24:41 crc kubenswrapper[4789]: E1008 14:24:41.002879 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593\": container with ID starting with 412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593 not found: ID does not exist" containerID="412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.002918 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593"} err="failed to get container status \"412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593\": rpc error: code = NotFound desc = could not find container \"412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593\": container with ID starting with 412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593 not found: ID does not exist" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.002948 4789 scope.go:117] "RemoveContainer" containerID="324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851" Oct 08 14:24:41 crc kubenswrapper[4789]: E1008 14:24:41.003304 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851\": container with ID starting with 324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851 not found: ID does not exist" containerID="324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.003334 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851"} err="failed to get container status \"324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851\": rpc error: code = NotFound desc = could not find container \"324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851\": container with ID starting with 324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851 not found: ID does not exist" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.003360 4789 scope.go:117] "RemoveContainer" containerID="412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.003557 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593"} err="failed to get container status \"412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593\": rpc error: code = NotFound desc = could not find container \"412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593\": container with ID starting with 412126abe26bc7ed4e7f278e4bf64dc2e7574d06db929065014d795e8c6c5593 not found: ID does not exist" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.003574 4789 scope.go:117] "RemoveContainer" containerID="324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.003747 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851"} err="failed to get container status \"324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851\": rpc error: code = NotFound desc = could not find container \"324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851\": container with ID starting with 324a803f9a32e112a0d5655c983cdfda9993dfbb857df60956f42751c06fb851 not found: ID does not exist" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.033180 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-config-data\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.033259 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lxxb\" (UniqueName: \"kubernetes.io/projected/a0d0eb62-71d0-4d90-90bf-046586854cf6-kube-api-access-8lxxb\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.033586 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.033676 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.033851 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d0eb62-71d0-4d90-90bf-046586854cf6-logs\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.135519 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d0eb62-71d0-4d90-90bf-046586854cf6-logs\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.135593 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-config-data\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.135637 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lxxb\" (UniqueName: \"kubernetes.io/projected/a0d0eb62-71d0-4d90-90bf-046586854cf6-kube-api-access-8lxxb\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.135766 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.135802 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.136547 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d0eb62-71d0-4d90-90bf-046586854cf6-logs\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.141542 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.141606 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-config-data\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.149458 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.154513 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lxxb\" (UniqueName: \"kubernetes.io/projected/a0d0eb62-71d0-4d90-90bf-046586854cf6-kube-api-access-8lxxb\") pod \"nova-metadata-0\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.292171 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:41 crc kubenswrapper[4789]: W1008 14:24:41.784000 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0d0eb62_71d0_4d90_90bf_046586854cf6.slice/crio-64f78f899d65bf60a63437882bcffec0f75b37cdf36763e24e85bab399d651e7 WatchSource:0}: Error finding container 64f78f899d65bf60a63437882bcffec0f75b37cdf36763e24e85bab399d651e7: Status 404 returned error can't find the container with id 64f78f899d65bf60a63437882bcffec0f75b37cdf36763e24e85bab399d651e7 Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.784068 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.860270 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 14:24:41 crc kubenswrapper[4789]: I1008 14:24:41.905108 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d0eb62-71d0-4d90-90bf-046586854cf6","Type":"ContainerStarted","Data":"64f78f899d65bf60a63437882bcffec0f75b37cdf36763e24e85bab399d651e7"} Oct 08 14:24:42 crc kubenswrapper[4789]: I1008 14:24:42.753359 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47f01d53-0d16-45cc-a6a7-f6ec90403969" path="/var/lib/kubelet/pods/47f01d53-0d16-45cc-a6a7-f6ec90403969/volumes" Oct 08 14:24:42 crc kubenswrapper[4789]: I1008 14:24:42.925789 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d0eb62-71d0-4d90-90bf-046586854cf6","Type":"ContainerStarted","Data":"ca581b859af141bd3b74181028f8be51e086d09b60a46876a0da0c1d4c16ba8b"} Oct 08 14:24:42 crc kubenswrapper[4789]: I1008 14:24:42.925833 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d0eb62-71d0-4d90-90bf-046586854cf6","Type":"ContainerStarted","Data":"f3b0624ff4e9bf2a1eabcc055368c6411157ea0b7f80ecb26bf60238681b483d"} Oct 08 14:24:42 crc kubenswrapper[4789]: I1008 14:24:42.943730 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.94370718 podStartE2EDuration="2.94370718s" podCreationTimestamp="2025-10-08 14:24:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:24:42.941560181 +0000 UTC m=+1422.848307673" watchObservedRunningTime="2025-10-08 14:24:42.94370718 +0000 UTC m=+1422.850454672" Oct 08 14:24:43 crc kubenswrapper[4789]: I1008 14:24:43.938060 4789 generic.go:334] "Generic (PLEG): container finished" podID="829d0843-9f3f-4bd4-b5e3-4562c7405f5f" containerID="508deddf75c87868a482535af0949a762abe0b0521a1144c6bb3ee5c019f6ca3" exitCode=0 Oct 08 14:24:43 crc kubenswrapper[4789]: I1008 14:24:43.938172 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v56bp" event={"ID":"829d0843-9f3f-4bd4-b5e3-4562c7405f5f","Type":"ContainerDied","Data":"508deddf75c87868a482535af0949a762abe0b0521a1144c6bb3ee5c019f6ca3"} Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.169270 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.169325 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.199165 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.272752 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.272823 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.461393 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.474909 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.545432 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bb6987789-gwcfk"] Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.545704 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" podUID="7cbf3178-8d77-47d9-beb0-804732748d97" containerName="dnsmasq-dns" containerID="cri-o://9e226dec43528f969019b70e47ed42a9dfda14f686a49c3629e8c726c49ae94f" gracePeriod=10 Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.967546 4789 generic.go:334] "Generic (PLEG): container finished" podID="7cbf3178-8d77-47d9-beb0-804732748d97" containerID="9e226dec43528f969019b70e47ed42a9dfda14f686a49c3629e8c726c49ae94f" exitCode=0 Oct 08 14:24:44 crc kubenswrapper[4789]: I1008 14:24:44.967604 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" event={"ID":"7cbf3178-8d77-47d9-beb0-804732748d97","Type":"ContainerDied","Data":"9e226dec43528f969019b70e47ed42a9dfda14f686a49c3629e8c726c49ae94f"} Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.030601 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.299936 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.363314 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.363680 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.421551 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-swift-storage-0\") pod \"7cbf3178-8d77-47d9-beb0-804732748d97\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.421645 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-config\") pod \"7cbf3178-8d77-47d9-beb0-804732748d97\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.421739 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-svc\") pod \"7cbf3178-8d77-47d9-beb0-804732748d97\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.421772 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-nb\") pod \"7cbf3178-8d77-47d9-beb0-804732748d97\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.421794 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-sb\") pod \"7cbf3178-8d77-47d9-beb0-804732748d97\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.421846 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chfd8\" (UniqueName: \"kubernetes.io/projected/7cbf3178-8d77-47d9-beb0-804732748d97-kube-api-access-chfd8\") pod \"7cbf3178-8d77-47d9-beb0-804732748d97\" (UID: \"7cbf3178-8d77-47d9-beb0-804732748d97\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.428613 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cbf3178-8d77-47d9-beb0-804732748d97-kube-api-access-chfd8" (OuterVolumeSpecName: "kube-api-access-chfd8") pod "7cbf3178-8d77-47d9-beb0-804732748d97" (UID: "7cbf3178-8d77-47d9-beb0-804732748d97"). InnerVolumeSpecName "kube-api-access-chfd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.480606 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7cbf3178-8d77-47d9-beb0-804732748d97" (UID: "7cbf3178-8d77-47d9-beb0-804732748d97"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.490056 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7cbf3178-8d77-47d9-beb0-804732748d97" (UID: "7cbf3178-8d77-47d9-beb0-804732748d97"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.490217 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7cbf3178-8d77-47d9-beb0-804732748d97" (UID: "7cbf3178-8d77-47d9-beb0-804732748d97"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.507614 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.510559 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7cbf3178-8d77-47d9-beb0-804732748d97" (UID: "7cbf3178-8d77-47d9-beb0-804732748d97"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.524221 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.524269 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.524282 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.524294 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.524307 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chfd8\" (UniqueName: \"kubernetes.io/projected/7cbf3178-8d77-47d9-beb0-804732748d97-kube-api-access-chfd8\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.618177 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-config" (OuterVolumeSpecName: "config") pod "7cbf3178-8d77-47d9-beb0-804732748d97" (UID: "7cbf3178-8d77-47d9-beb0-804732748d97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.625343 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b9n7\" (UniqueName: \"kubernetes.io/projected/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-kube-api-access-9b9n7\") pod \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.625423 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-combined-ca-bundle\") pod \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.625615 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-scripts\") pod \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.625748 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-config-data\") pod \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\" (UID: \"829d0843-9f3f-4bd4-b5e3-4562c7405f5f\") " Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.626219 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cbf3178-8d77-47d9-beb0-804732748d97-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.629854 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-scripts" (OuterVolumeSpecName: "scripts") pod "829d0843-9f3f-4bd4-b5e3-4562c7405f5f" (UID: "829d0843-9f3f-4bd4-b5e3-4562c7405f5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.631565 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-kube-api-access-9b9n7" (OuterVolumeSpecName: "kube-api-access-9b9n7") pod "829d0843-9f3f-4bd4-b5e3-4562c7405f5f" (UID: "829d0843-9f3f-4bd4-b5e3-4562c7405f5f"). InnerVolumeSpecName "kube-api-access-9b9n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.671587 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-config-data" (OuterVolumeSpecName: "config-data") pod "829d0843-9f3f-4bd4-b5e3-4562c7405f5f" (UID: "829d0843-9f3f-4bd4-b5e3-4562c7405f5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.677401 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "829d0843-9f3f-4bd4-b5e3-4562c7405f5f" (UID: "829d0843-9f3f-4bd4-b5e3-4562c7405f5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.728502 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.728832 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.728925 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b9n7\" (UniqueName: \"kubernetes.io/projected/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-kube-api-access-9b9n7\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.729024 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829d0843-9f3f-4bd4-b5e3-4562c7405f5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.832540 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.833178 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="a4509515-2cf4-40a1-9873-588155ad596e" containerName="kube-state-metrics" containerID="cri-o://459318a2a91161308a2cbf8b5312ea5c90022b0c88f09a77a38c25c7ac3506a8" gracePeriod=30 Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.979547 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-v56bp" event={"ID":"829d0843-9f3f-4bd4-b5e3-4562c7405f5f","Type":"ContainerDied","Data":"29d3a3c8986848772d6650b2071d0b3c9d56b9e5d9731b97e257c55359099afa"} Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.979604 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29d3a3c8986848772d6650b2071d0b3c9d56b9e5d9731b97e257c55359099afa" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.980794 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-v56bp" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.982770 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" event={"ID":"7cbf3178-8d77-47d9-beb0-804732748d97","Type":"ContainerDied","Data":"2fefad0d6e111c73573721cdbf78c238071eb4dd4b9546dcff761034bb77165e"} Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.982799 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb6987789-gwcfk" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.982832 4789 scope.go:117] "RemoveContainer" containerID="9e226dec43528f969019b70e47ed42a9dfda14f686a49c3629e8c726c49ae94f" Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.984748 4789 generic.go:334] "Generic (PLEG): container finished" podID="a4509515-2cf4-40a1-9873-588155ad596e" containerID="459318a2a91161308a2cbf8b5312ea5c90022b0c88f09a77a38c25c7ac3506a8" exitCode=2 Oct 08 14:24:45 crc kubenswrapper[4789]: I1008 14:24:45.984829 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a4509515-2cf4-40a1-9873-588155ad596e","Type":"ContainerDied","Data":"459318a2a91161308a2cbf8b5312ea5c90022b0c88f09a77a38c25c7ac3506a8"} Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.041095 4789 scope.go:117] "RemoveContainer" containerID="b8e92044f3c94beedb36cb345ffadc3826d651bb3ed186b182c22681d72f6cd2" Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.046654 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bb6987789-gwcfk"] Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.060108 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bb6987789-gwcfk"] Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.167746 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.168299 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-api" containerID="cri-o://d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98" gracePeriod=30 Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.173057 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-log" containerID="cri-o://4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311" gracePeriod=30 Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.216030 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.234836 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.235122 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerName="nova-metadata-log" containerID="cri-o://f3b0624ff4e9bf2a1eabcc055368c6411157ea0b7f80ecb26bf60238681b483d" gracePeriod=30 Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.235252 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerName="nova-metadata-metadata" containerID="cri-o://ca581b859af141bd3b74181028f8be51e086d09b60a46876a0da0c1d4c16ba8b" gracePeriod=30 Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.296115 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.296191 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.515689 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.653590 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtxkm\" (UniqueName: \"kubernetes.io/projected/a4509515-2cf4-40a1-9873-588155ad596e-kube-api-access-jtxkm\") pod \"a4509515-2cf4-40a1-9873-588155ad596e\" (UID: \"a4509515-2cf4-40a1-9873-588155ad596e\") " Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.664286 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4509515-2cf4-40a1-9873-588155ad596e-kube-api-access-jtxkm" (OuterVolumeSpecName: "kube-api-access-jtxkm") pod "a4509515-2cf4-40a1-9873-588155ad596e" (UID: "a4509515-2cf4-40a1-9873-588155ad596e"). InnerVolumeSpecName "kube-api-access-jtxkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.750293 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cbf3178-8d77-47d9-beb0-804732748d97" path="/var/lib/kubelet/pods/7cbf3178-8d77-47d9-beb0-804732748d97/volumes" Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.759596 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtxkm\" (UniqueName: \"kubernetes.io/projected/a4509515-2cf4-40a1-9873-588155ad596e-kube-api-access-jtxkm\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.996548 4789 generic.go:334] "Generic (PLEG): container finished" podID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerID="ca581b859af141bd3b74181028f8be51e086d09b60a46876a0da0c1d4c16ba8b" exitCode=0 Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.996607 4789 generic.go:334] "Generic (PLEG): container finished" podID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerID="f3b0624ff4e9bf2a1eabcc055368c6411157ea0b7f80ecb26bf60238681b483d" exitCode=143 Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.996632 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d0eb62-71d0-4d90-90bf-046586854cf6","Type":"ContainerDied","Data":"ca581b859af141bd3b74181028f8be51e086d09b60a46876a0da0c1d4c16ba8b"} Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.996673 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d0eb62-71d0-4d90-90bf-046586854cf6","Type":"ContainerDied","Data":"f3b0624ff4e9bf2a1eabcc055368c6411157ea0b7f80ecb26bf60238681b483d"} Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.996689 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0d0eb62-71d0-4d90-90bf-046586854cf6","Type":"ContainerDied","Data":"64f78f899d65bf60a63437882bcffec0f75b37cdf36763e24e85bab399d651e7"} Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.996705 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64f78f899d65bf60a63437882bcffec0f75b37cdf36763e24e85bab399d651e7" Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.999675 4789 generic.go:334] "Generic (PLEG): container finished" podID="88bebca2-d7f3-436d-b034-69579813cbc9" containerID="4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311" exitCode=143 Oct 08 14:24:46 crc kubenswrapper[4789]: I1008 14:24:46.999736 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88bebca2-d7f3-436d-b034-69579813cbc9","Type":"ContainerDied","Data":"4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311"} Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.001104 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.001105 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a4509515-2cf4-40a1-9873-588155ad596e","Type":"ContainerDied","Data":"4372a74793ba3a1d74029c158d659c03ff293c78f7baff55f03aae6375ae5748"} Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.001191 4789 scope.go:117] "RemoveContainer" containerID="459318a2a91161308a2cbf8b5312ea5c90022b0c88f09a77a38c25c7ac3506a8" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.001243 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c03e9357-b2ad-421e-96a7-d513673139a1" containerName="nova-scheduler-scheduler" containerID="cri-o://fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96" gracePeriod=30 Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.034413 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.046683 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.054859 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.099828 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 14:24:47 crc kubenswrapper[4789]: E1008 14:24:47.100411 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerName="nova-metadata-metadata" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100437 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerName="nova-metadata-metadata" Oct 08 14:24:47 crc kubenswrapper[4789]: E1008 14:24:47.100467 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4509515-2cf4-40a1-9873-588155ad596e" containerName="kube-state-metrics" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100476 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4509515-2cf4-40a1-9873-588155ad596e" containerName="kube-state-metrics" Oct 08 14:24:47 crc kubenswrapper[4789]: E1008 14:24:47.100503 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cbf3178-8d77-47d9-beb0-804732748d97" containerName="init" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100511 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cbf3178-8d77-47d9-beb0-804732748d97" containerName="init" Oct 08 14:24:47 crc kubenswrapper[4789]: E1008 14:24:47.100530 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerName="nova-metadata-log" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100538 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerName="nova-metadata-log" Oct 08 14:24:47 crc kubenswrapper[4789]: E1008 14:24:47.100549 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829d0843-9f3f-4bd4-b5e3-4562c7405f5f" containerName="nova-manage" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100557 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="829d0843-9f3f-4bd4-b5e3-4562c7405f5f" containerName="nova-manage" Oct 08 14:24:47 crc kubenswrapper[4789]: E1008 14:24:47.100578 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cbf3178-8d77-47d9-beb0-804732748d97" containerName="dnsmasq-dns" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100586 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cbf3178-8d77-47d9-beb0-804732748d97" containerName="dnsmasq-dns" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100809 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4509515-2cf4-40a1-9873-588155ad596e" containerName="kube-state-metrics" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100842 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="829d0843-9f3f-4bd4-b5e3-4562c7405f5f" containerName="nova-manage" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100859 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cbf3178-8d77-47d9-beb0-804732748d97" containerName="dnsmasq-dns" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100884 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerName="nova-metadata-metadata" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.100897 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d0eb62-71d0-4d90-90bf-046586854cf6" containerName="nova-metadata-log" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.101701 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.104491 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.104736 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.109148 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.164688 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-config-data\") pod \"a0d0eb62-71d0-4d90-90bf-046586854cf6\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.165151 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-nova-metadata-tls-certs\") pod \"a0d0eb62-71d0-4d90-90bf-046586854cf6\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.165298 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-combined-ca-bundle\") pod \"a0d0eb62-71d0-4d90-90bf-046586854cf6\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.165426 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lxxb\" (UniqueName: \"kubernetes.io/projected/a0d0eb62-71d0-4d90-90bf-046586854cf6-kube-api-access-8lxxb\") pod \"a0d0eb62-71d0-4d90-90bf-046586854cf6\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.165482 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d0eb62-71d0-4d90-90bf-046586854cf6-logs\") pod \"a0d0eb62-71d0-4d90-90bf-046586854cf6\" (UID: \"a0d0eb62-71d0-4d90-90bf-046586854cf6\") " Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.167015 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0d0eb62-71d0-4d90-90bf-046586854cf6-logs" (OuterVolumeSpecName: "logs") pod "a0d0eb62-71d0-4d90-90bf-046586854cf6" (UID: "a0d0eb62-71d0-4d90-90bf-046586854cf6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.178535 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0d0eb62-71d0-4d90-90bf-046586854cf6-kube-api-access-8lxxb" (OuterVolumeSpecName: "kube-api-access-8lxxb") pod "a0d0eb62-71d0-4d90-90bf-046586854cf6" (UID: "a0d0eb62-71d0-4d90-90bf-046586854cf6"). InnerVolumeSpecName "kube-api-access-8lxxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.209445 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0d0eb62-71d0-4d90-90bf-046586854cf6" (UID: "a0d0eb62-71d0-4d90-90bf-046586854cf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.222225 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-config-data" (OuterVolumeSpecName: "config-data") pod "a0d0eb62-71d0-4d90-90bf-046586854cf6" (UID: "a0d0eb62-71d0-4d90-90bf-046586854cf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.249329 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a0d0eb62-71d0-4d90-90bf-046586854cf6" (UID: "a0d0eb62-71d0-4d90-90bf-046586854cf6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.268913 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/08e907c7-3b60-4d50-977c-32d05a3ce80f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.269333 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08e907c7-3b60-4d50-977c-32d05a3ce80f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.269463 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b28j7\" (UniqueName: \"kubernetes.io/projected/08e907c7-3b60-4d50-977c-32d05a3ce80f-kube-api-access-b28j7\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.269573 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/08e907c7-3b60-4d50-977c-32d05a3ce80f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.269721 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.269794 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lxxb\" (UniqueName: \"kubernetes.io/projected/a0d0eb62-71d0-4d90-90bf-046586854cf6-kube-api-access-8lxxb\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.269871 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d0eb62-71d0-4d90-90bf-046586854cf6-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.269944 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.270070 4789 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d0eb62-71d0-4d90-90bf-046586854cf6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.371731 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08e907c7-3b60-4d50-977c-32d05a3ce80f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.372114 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b28j7\" (UniqueName: \"kubernetes.io/projected/08e907c7-3b60-4d50-977c-32d05a3ce80f-kube-api-access-b28j7\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.372258 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/08e907c7-3b60-4d50-977c-32d05a3ce80f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.372398 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/08e907c7-3b60-4d50-977c-32d05a3ce80f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.376801 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08e907c7-3b60-4d50-977c-32d05a3ce80f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.376813 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/08e907c7-3b60-4d50-977c-32d05a3ce80f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.383963 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/08e907c7-3b60-4d50-977c-32d05a3ce80f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.394661 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b28j7\" (UniqueName: \"kubernetes.io/projected/08e907c7-3b60-4d50-977c-32d05a3ce80f-kube-api-access-b28j7\") pod \"kube-state-metrics-0\" (UID: \"08e907c7-3b60-4d50-977c-32d05a3ce80f\") " pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.421139 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 14:24:47 crc kubenswrapper[4789]: I1008 14:24:47.888018 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 14:24:47 crc kubenswrapper[4789]: W1008 14:24:47.892319 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08e907c7_3b60_4d50_977c_32d05a3ce80f.slice/crio-db65cf654ff39de3da1a53578eeb109bdb774e2ad8d4c14d1cff7e650b29ee4c WatchSource:0}: Error finding container db65cf654ff39de3da1a53578eeb109bdb774e2ad8d4c14d1cff7e650b29ee4c: Status 404 returned error can't find the container with id db65cf654ff39de3da1a53578eeb109bdb774e2ad8d4c14d1cff7e650b29ee4c Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.012926 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.012952 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"08e907c7-3b60-4d50-977c-32d05a3ce80f","Type":"ContainerStarted","Data":"db65cf654ff39de3da1a53578eeb109bdb774e2ad8d4c14d1cff7e650b29ee4c"} Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.092337 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.108832 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.117718 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.120216 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.123304 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.123737 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.125696 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.278832 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.279444 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="ceilometer-central-agent" containerID="cri-o://21ec477643d66c1646db0374733fea8d174901474abbc958fff28425b0b3d3ee" gracePeriod=30 Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.279666 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="proxy-httpd" containerID="cri-o://461690d543781b3c72717e5cc6694c5a7c839c5e48b71e11ee21c99d3e03b30a" gracePeriod=30 Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.279815 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="sg-core" containerID="cri-o://0d52a25246184153286f994395ebb0b5b8722d0f40c5743dc54a7c238d1e5f3b" gracePeriod=30 Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.279981 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="ceilometer-notification-agent" containerID="cri-o://1d0969b22c0457046963c721f2fe13930d401fb05ce2543511f2886ebca4be00" gracePeriod=30 Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.291388 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/426d4e55-0509-4936-bede-51b8dc8ccc90-logs\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.291502 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.291576 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-config-data\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.291641 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/426d4e55-0509-4936-bede-51b8dc8ccc90-kube-api-access-vnnt5\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.291668 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.393435 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/426d4e55-0509-4936-bede-51b8dc8ccc90-logs\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.393686 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.393733 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-config-data\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.393775 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/426d4e55-0509-4936-bede-51b8dc8ccc90-kube-api-access-vnnt5\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.393793 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.394063 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/426d4e55-0509-4936-bede-51b8dc8ccc90-logs\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.398903 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.404617 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-config-data\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.405480 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.419831 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/426d4e55-0509-4936-bede-51b8dc8ccc90-kube-api-access-vnnt5\") pod \"nova-metadata-0\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.442180 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.741718 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0d0eb62-71d0-4d90-90bf-046586854cf6" path="/var/lib/kubelet/pods/a0d0eb62-71d0-4d90-90bf-046586854cf6/volumes" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.742590 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4509515-2cf4-40a1-9873-588155ad596e" path="/var/lib/kubelet/pods/a4509515-2cf4-40a1-9873-588155ad596e/volumes" Oct 08 14:24:48 crc kubenswrapper[4789]: I1008 14:24:48.940367 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.031311 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerID="461690d543781b3c72717e5cc6694c5a7c839c5e48b71e11ee21c99d3e03b30a" exitCode=0 Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.031341 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerID="0d52a25246184153286f994395ebb0b5b8722d0f40c5743dc54a7c238d1e5f3b" exitCode=2 Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.031350 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerID="21ec477643d66c1646db0374733fea8d174901474abbc958fff28425b0b3d3ee" exitCode=0 Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.031380 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerDied","Data":"461690d543781b3c72717e5cc6694c5a7c839c5e48b71e11ee21c99d3e03b30a"} Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.031414 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerDied","Data":"0d52a25246184153286f994395ebb0b5b8722d0f40c5743dc54a7c238d1e5f3b"} Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.031426 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerDied","Data":"21ec477643d66c1646db0374733fea8d174901474abbc958fff28425b0b3d3ee"} Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.034287 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"08e907c7-3b60-4d50-977c-32d05a3ce80f","Type":"ContainerStarted","Data":"bed409ac8f933a51822d0e467e470631ae6fbe31d4a98b091f6fe0cb61271804"} Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.034417 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.036900 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"426d4e55-0509-4936-bede-51b8dc8ccc90","Type":"ContainerStarted","Data":"7f19fe596baf3ca80d468c9d5ca2a7a29668ce99f713f90ae1583ac4d1d2e66c"} Oct 08 14:24:49 crc kubenswrapper[4789]: I1008 14:24:49.053747 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.691409545 podStartE2EDuration="2.053727633s" podCreationTimestamp="2025-10-08 14:24:47 +0000 UTC" firstStartedPulling="2025-10-08 14:24:47.895813958 +0000 UTC m=+1427.802561450" lastFinishedPulling="2025-10-08 14:24:48.258132046 +0000 UTC m=+1428.164879538" observedRunningTime="2025-10-08 14:24:49.052214262 +0000 UTC m=+1428.958961754" watchObservedRunningTime="2025-10-08 14:24:49.053727633 +0000 UTC m=+1428.960475125" Oct 08 14:24:49 crc kubenswrapper[4789]: E1008 14:24:49.173632 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 14:24:49 crc kubenswrapper[4789]: E1008 14:24:49.178354 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 14:24:49 crc kubenswrapper[4789]: E1008 14:24:49.179777 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 14:24:49 crc kubenswrapper[4789]: E1008 14:24:49.179832 4789 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c03e9357-b2ad-421e-96a7-d513673139a1" containerName="nova-scheduler-scheduler" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.046485 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"426d4e55-0509-4936-bede-51b8dc8ccc90","Type":"ContainerStarted","Data":"c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c"} Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.047743 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"426d4e55-0509-4936-bede-51b8dc8ccc90","Type":"ContainerStarted","Data":"74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190"} Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.050422 4789 generic.go:334] "Generic (PLEG): container finished" podID="4aa68379-349e-48f5-b6c5-d2b694639e9f" containerID="ccbe308e2801588c3627cf00337a6708ad65f29cc86ef71ba89a14e63d4ee497" exitCode=0 Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.050458 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dr8h4" event={"ID":"4aa68379-349e-48f5-b6c5-d2b694639e9f","Type":"ContainerDied","Data":"ccbe308e2801588c3627cf00337a6708ad65f29cc86ef71ba89a14e63d4ee497"} Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.067347 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.067321379 podStartE2EDuration="2.067321379s" podCreationTimestamp="2025-10-08 14:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:24:50.063399782 +0000 UTC m=+1429.970147274" watchObservedRunningTime="2025-10-08 14:24:50.067321379 +0000 UTC m=+1429.974068871" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.558686 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.642389 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kh4n\" (UniqueName: \"kubernetes.io/projected/88bebca2-d7f3-436d-b034-69579813cbc9-kube-api-access-2kh4n\") pod \"88bebca2-d7f3-436d-b034-69579813cbc9\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.642640 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-combined-ca-bundle\") pod \"88bebca2-d7f3-436d-b034-69579813cbc9\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.642712 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-config-data\") pod \"88bebca2-d7f3-436d-b034-69579813cbc9\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.642790 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88bebca2-d7f3-436d-b034-69579813cbc9-logs\") pod \"88bebca2-d7f3-436d-b034-69579813cbc9\" (UID: \"88bebca2-d7f3-436d-b034-69579813cbc9\") " Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.643580 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88bebca2-d7f3-436d-b034-69579813cbc9-logs" (OuterVolumeSpecName: "logs") pod "88bebca2-d7f3-436d-b034-69579813cbc9" (UID: "88bebca2-d7f3-436d-b034-69579813cbc9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.650172 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88bebca2-d7f3-436d-b034-69579813cbc9-kube-api-access-2kh4n" (OuterVolumeSpecName: "kube-api-access-2kh4n") pod "88bebca2-d7f3-436d-b034-69579813cbc9" (UID: "88bebca2-d7f3-436d-b034-69579813cbc9"). InnerVolumeSpecName "kube-api-access-2kh4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.680142 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88bebca2-d7f3-436d-b034-69579813cbc9" (UID: "88bebca2-d7f3-436d-b034-69579813cbc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.682932 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-config-data" (OuterVolumeSpecName: "config-data") pod "88bebca2-d7f3-436d-b034-69579813cbc9" (UID: "88bebca2-d7f3-436d-b034-69579813cbc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.745132 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.745166 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88bebca2-d7f3-436d-b034-69579813cbc9-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.745176 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88bebca2-d7f3-436d-b034-69579813cbc9-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:50 crc kubenswrapper[4789]: I1008 14:24:50.745185 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kh4n\" (UniqueName: \"kubernetes.io/projected/88bebca2-d7f3-436d-b034-69579813cbc9-kube-api-access-2kh4n\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.060748 4789 generic.go:334] "Generic (PLEG): container finished" podID="88bebca2-d7f3-436d-b034-69579813cbc9" containerID="d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98" exitCode=0 Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.060814 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.060847 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88bebca2-d7f3-436d-b034-69579813cbc9","Type":"ContainerDied","Data":"d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98"} Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.060874 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88bebca2-d7f3-436d-b034-69579813cbc9","Type":"ContainerDied","Data":"91985a24003611da96a389a0236bdd5858c928351c8b2e4b420a2304e7baf244"} Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.060889 4789 scope.go:117] "RemoveContainer" containerID="d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.087129 4789 scope.go:117] "RemoveContainer" containerID="4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.104454 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.139928 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.165141 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 14:24:51 crc kubenswrapper[4789]: E1008 14:24:51.165875 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-log" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.165926 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-log" Oct 08 14:24:51 crc kubenswrapper[4789]: E1008 14:24:51.165962 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-api" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.165971 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-api" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.166312 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-log" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.166369 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" containerName="nova-api-api" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.167488 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.169407 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.174888 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.189675 4789 scope.go:117] "RemoveContainer" containerID="d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98" Oct 08 14:24:51 crc kubenswrapper[4789]: E1008 14:24:51.191906 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98\": container with ID starting with d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98 not found: ID does not exist" containerID="d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.191956 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98"} err="failed to get container status \"d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98\": rpc error: code = NotFound desc = could not find container \"d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98\": container with ID starting with d4a503903b0f5cd3469040fd868cfd4c1ff7c98eb4b4b952f4de26e15cb3ac98 not found: ID does not exist" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.192005 4789 scope.go:117] "RemoveContainer" containerID="4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311" Oct 08 14:24:51 crc kubenswrapper[4789]: E1008 14:24:51.192413 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311\": container with ID starting with 4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311 not found: ID does not exist" containerID="4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.192457 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311"} err="failed to get container status \"4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311\": rpc error: code = NotFound desc = could not find container \"4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311\": container with ID starting with 4945448cf93b8d0b8a1bf4e7c235a08144496d289e49e13f126691e09dc6d311 not found: ID does not exist" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.255050 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrdbm\" (UniqueName: \"kubernetes.io/projected/cb208d86-8b8b-4114-8638-7d8a98a6223f-kube-api-access-vrdbm\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.255113 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-config-data\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.255210 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb208d86-8b8b-4114-8638-7d8a98a6223f-logs\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.255236 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.357509 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrdbm\" (UniqueName: \"kubernetes.io/projected/cb208d86-8b8b-4114-8638-7d8a98a6223f-kube-api-access-vrdbm\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.357561 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-config-data\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.357639 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb208d86-8b8b-4114-8638-7d8a98a6223f-logs\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.357661 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.360167 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb208d86-8b8b-4114-8638-7d8a98a6223f-logs\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.367189 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-config-data\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.367292 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.375219 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrdbm\" (UniqueName: \"kubernetes.io/projected/cb208d86-8b8b-4114-8638-7d8a98a6223f-kube-api-access-vrdbm\") pod \"nova-api-0\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.512314 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.521039 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.528337 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.665592 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-combined-ca-bundle\") pod \"4aa68379-349e-48f5-b6c5-d2b694639e9f\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.665703 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-combined-ca-bundle\") pod \"c03e9357-b2ad-421e-96a7-d513673139a1\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.665792 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf8rv\" (UniqueName: \"kubernetes.io/projected/c03e9357-b2ad-421e-96a7-d513673139a1-kube-api-access-tf8rv\") pod \"c03e9357-b2ad-421e-96a7-d513673139a1\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.665829 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-config-data\") pod \"c03e9357-b2ad-421e-96a7-d513673139a1\" (UID: \"c03e9357-b2ad-421e-96a7-d513673139a1\") " Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.665879 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzb96\" (UniqueName: \"kubernetes.io/projected/4aa68379-349e-48f5-b6c5-d2b694639e9f-kube-api-access-qzb96\") pod \"4aa68379-349e-48f5-b6c5-d2b694639e9f\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.665919 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-scripts\") pod \"4aa68379-349e-48f5-b6c5-d2b694639e9f\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.665943 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-config-data\") pod \"4aa68379-349e-48f5-b6c5-d2b694639e9f\" (UID: \"4aa68379-349e-48f5-b6c5-d2b694639e9f\") " Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.671338 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa68379-349e-48f5-b6c5-d2b694639e9f-kube-api-access-qzb96" (OuterVolumeSpecName: "kube-api-access-qzb96") pod "4aa68379-349e-48f5-b6c5-d2b694639e9f" (UID: "4aa68379-349e-48f5-b6c5-d2b694639e9f"). InnerVolumeSpecName "kube-api-access-qzb96". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.674294 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03e9357-b2ad-421e-96a7-d513673139a1-kube-api-access-tf8rv" (OuterVolumeSpecName: "kube-api-access-tf8rv") pod "c03e9357-b2ad-421e-96a7-d513673139a1" (UID: "c03e9357-b2ad-421e-96a7-d513673139a1"). InnerVolumeSpecName "kube-api-access-tf8rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.674497 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-scripts" (OuterVolumeSpecName: "scripts") pod "4aa68379-349e-48f5-b6c5-d2b694639e9f" (UID: "4aa68379-349e-48f5-b6c5-d2b694639e9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.694330 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4aa68379-349e-48f5-b6c5-d2b694639e9f" (UID: "4aa68379-349e-48f5-b6c5-d2b694639e9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.700321 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-config-data" (OuterVolumeSpecName: "config-data") pod "c03e9357-b2ad-421e-96a7-d513673139a1" (UID: "c03e9357-b2ad-421e-96a7-d513673139a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.703145 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c03e9357-b2ad-421e-96a7-d513673139a1" (UID: "c03e9357-b2ad-421e-96a7-d513673139a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.711623 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-config-data" (OuterVolumeSpecName: "config-data") pod "4aa68379-349e-48f5-b6c5-d2b694639e9f" (UID: "4aa68379-349e-48f5-b6c5-d2b694639e9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.769069 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.769113 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.769126 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf8rv\" (UniqueName: \"kubernetes.io/projected/c03e9357-b2ad-421e-96a7-d513673139a1-kube-api-access-tf8rv\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.769141 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c03e9357-b2ad-421e-96a7-d513673139a1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.769154 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzb96\" (UniqueName: \"kubernetes.io/projected/4aa68379-349e-48f5-b6c5-d2b694639e9f-kube-api-access-qzb96\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.769165 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:51 crc kubenswrapper[4789]: I1008 14:24:51.769176 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aa68379-349e-48f5-b6c5-d2b694639e9f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.016575 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:24:52 crc kubenswrapper[4789]: W1008 14:24:52.021692 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb208d86_8b8b_4114_8638_7d8a98a6223f.slice/crio-8ff668821e7a7ad66dab2b17d6c619267cebd0bde0fc9783059f39d2ec5ee224 WatchSource:0}: Error finding container 8ff668821e7a7ad66dab2b17d6c619267cebd0bde0fc9783059f39d2ec5ee224: Status 404 returned error can't find the container with id 8ff668821e7a7ad66dab2b17d6c619267cebd0bde0fc9783059f39d2ec5ee224 Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.075407 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dr8h4" event={"ID":"4aa68379-349e-48f5-b6c5-d2b694639e9f","Type":"ContainerDied","Data":"a3cb9561eb4726068aa7b7ea4e07b8f551e6308e2bfe71dba5989bbc71271930"} Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.075490 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3cb9561eb4726068aa7b7ea4e07b8f551e6308e2bfe71dba5989bbc71271930" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.075667 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dr8h4" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.081056 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb208d86-8b8b-4114-8638-7d8a98a6223f","Type":"ContainerStarted","Data":"8ff668821e7a7ad66dab2b17d6c619267cebd0bde0fc9783059f39d2ec5ee224"} Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.087282 4789 generic.go:334] "Generic (PLEG): container finished" podID="c03e9357-b2ad-421e-96a7-d513673139a1" containerID="fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96" exitCode=0 Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.087805 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.088500 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c03e9357-b2ad-421e-96a7-d513673139a1","Type":"ContainerDied","Data":"fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96"} Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.088530 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c03e9357-b2ad-421e-96a7-d513673139a1","Type":"ContainerDied","Data":"447961291e5f556e48f377196d497a70c0ff97ae6911fbe33ed084f5b384279e"} Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.088550 4789 scope.go:117] "RemoveContainer" containerID="fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.136584 4789 scope.go:117] "RemoveContainer" containerID="fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96" Oct 08 14:24:52 crc kubenswrapper[4789]: E1008 14:24:52.137846 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96\": container with ID starting with fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96 not found: ID does not exist" containerID="fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.137892 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96"} err="failed to get container status \"fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96\": rpc error: code = NotFound desc = could not find container \"fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96\": container with ID starting with fa488480c53161049df08ff5b8727945093715ba241eb29f7cb862b9c6c36b96 not found: ID does not exist" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.170733 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 14:24:52 crc kubenswrapper[4789]: E1008 14:24:52.171315 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c03e9357-b2ad-421e-96a7-d513673139a1" containerName="nova-scheduler-scheduler" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.171331 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c03e9357-b2ad-421e-96a7-d513673139a1" containerName="nova-scheduler-scheduler" Oct 08 14:24:52 crc kubenswrapper[4789]: E1008 14:24:52.171361 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa68379-349e-48f5-b6c5-d2b694639e9f" containerName="nova-cell1-conductor-db-sync" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.171369 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa68379-349e-48f5-b6c5-d2b694639e9f" containerName="nova-cell1-conductor-db-sync" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.171598 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa68379-349e-48f5-b6c5-d2b694639e9f" containerName="nova-cell1-conductor-db-sync" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.171630 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c03e9357-b2ad-421e-96a7-d513673139a1" containerName="nova-scheduler-scheduler" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.172518 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.181000 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.190806 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.207067 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.221251 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.238699 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.240268 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.249243 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.250350 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.292728 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af59a4db-7cb6-4e1f-8381-43e0f13e5875-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"af59a4db-7cb6-4e1f-8381-43e0f13e5875\") " pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.292840 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af59a4db-7cb6-4e1f-8381-43e0f13e5875-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"af59a4db-7cb6-4e1f-8381-43e0f13e5875\") " pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.292864 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn6d2\" (UniqueName: \"kubernetes.io/projected/af59a4db-7cb6-4e1f-8381-43e0f13e5875-kube-api-access-tn6d2\") pod \"nova-cell1-conductor-0\" (UID: \"af59a4db-7cb6-4e1f-8381-43e0f13e5875\") " pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.394829 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.394903 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af59a4db-7cb6-4e1f-8381-43e0f13e5875-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"af59a4db-7cb6-4e1f-8381-43e0f13e5875\") " pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.394923 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn6d2\" (UniqueName: \"kubernetes.io/projected/af59a4db-7cb6-4e1f-8381-43e0f13e5875-kube-api-access-tn6d2\") pod \"nova-cell1-conductor-0\" (UID: \"af59a4db-7cb6-4e1f-8381-43e0f13e5875\") " pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.394961 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw795\" (UniqueName: \"kubernetes.io/projected/bcb5bfe8-89a1-4085-a486-cce5e4e56789-kube-api-access-sw795\") pod \"nova-scheduler-0\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.395081 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af59a4db-7cb6-4e1f-8381-43e0f13e5875-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"af59a4db-7cb6-4e1f-8381-43e0f13e5875\") " pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.395103 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-config-data\") pod \"nova-scheduler-0\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.401882 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af59a4db-7cb6-4e1f-8381-43e0f13e5875-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"af59a4db-7cb6-4e1f-8381-43e0f13e5875\") " pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.405067 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af59a4db-7cb6-4e1f-8381-43e0f13e5875-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"af59a4db-7cb6-4e1f-8381-43e0f13e5875\") " pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.413570 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn6d2\" (UniqueName: \"kubernetes.io/projected/af59a4db-7cb6-4e1f-8381-43e0f13e5875-kube-api-access-tn6d2\") pod \"nova-cell1-conductor-0\" (UID: \"af59a4db-7cb6-4e1f-8381-43e0f13e5875\") " pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.497221 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-config-data\") pod \"nova-scheduler-0\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.497607 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.497787 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw795\" (UniqueName: \"kubernetes.io/projected/bcb5bfe8-89a1-4085-a486-cce5e4e56789-kube-api-access-sw795\") pod \"nova-scheduler-0\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.501688 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.502921 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-config-data\") pod \"nova-scheduler-0\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.519497 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw795\" (UniqueName: \"kubernetes.io/projected/bcb5bfe8-89a1-4085-a486-cce5e4e56789-kube-api-access-sw795\") pod \"nova-scheduler-0\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.536544 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.572423 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.808153 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88bebca2-d7f3-436d-b034-69579813cbc9" path="/var/lib/kubelet/pods/88bebca2-d7f3-436d-b034-69579813cbc9/volumes" Oct 08 14:24:52 crc kubenswrapper[4789]: I1008 14:24:52.809313 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03e9357-b2ad-421e-96a7-d513673139a1" path="/var/lib/kubelet/pods/c03e9357-b2ad-421e-96a7-d513673139a1/volumes" Oct 08 14:24:53 crc kubenswrapper[4789]: I1008 14:24:53.065024 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 14:24:53 crc kubenswrapper[4789]: I1008 14:24:53.087825 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:24:53 crc kubenswrapper[4789]: W1008 14:24:53.100211 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcb5bfe8_89a1_4085_a486_cce5e4e56789.slice/crio-3fc6b03bf564dcc0df56c3f69c99dc70645bfef18a6736c3ac35f42547e1cecf WatchSource:0}: Error finding container 3fc6b03bf564dcc0df56c3f69c99dc70645bfef18a6736c3ac35f42547e1cecf: Status 404 returned error can't find the container with id 3fc6b03bf564dcc0df56c3f69c99dc70645bfef18a6736c3ac35f42547e1cecf Oct 08 14:24:53 crc kubenswrapper[4789]: I1008 14:24:53.114771 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb208d86-8b8b-4114-8638-7d8a98a6223f","Type":"ContainerStarted","Data":"f4756dad6ac88bdf861036d865cbae522a4d123bb70d7a18a08a318a35635fe3"} Oct 08 14:24:53 crc kubenswrapper[4789]: I1008 14:24:53.114818 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb208d86-8b8b-4114-8638-7d8a98a6223f","Type":"ContainerStarted","Data":"98749808ef5b82555ba4234bb16e54e5ae68c689df50f752330e6b330c7cc4e1"} Oct 08 14:24:53 crc kubenswrapper[4789]: I1008 14:24:53.124317 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"af59a4db-7cb6-4e1f-8381-43e0f13e5875","Type":"ContainerStarted","Data":"d3d5c8228d382560868f93a98b420ca2ee3bd770b80be34d2f1f64efcb656a52"} Oct 08 14:24:53 crc kubenswrapper[4789]: I1008 14:24:53.149225 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.149207059 podStartE2EDuration="2.149207059s" podCreationTimestamp="2025-10-08 14:24:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:24:53.135552959 +0000 UTC m=+1433.042300451" watchObservedRunningTime="2025-10-08 14:24:53.149207059 +0000 UTC m=+1433.055954551" Oct 08 14:24:53 crc kubenswrapper[4789]: I1008 14:24:53.443394 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 14:24:53 crc kubenswrapper[4789]: I1008 14:24:53.444712 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 14:24:54 crc kubenswrapper[4789]: I1008 14:24:54.138666 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"af59a4db-7cb6-4e1f-8381-43e0f13e5875","Type":"ContainerStarted","Data":"146a8315f04637134bd65f74b07437cf192feee1e100a1de353b70b0c7497bde"} Oct 08 14:24:54 crc kubenswrapper[4789]: I1008 14:24:54.139719 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 08 14:24:54 crc kubenswrapper[4789]: I1008 14:24:54.141587 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bcb5bfe8-89a1-4085-a486-cce5e4e56789","Type":"ContainerStarted","Data":"f34a8ac20d6f6dd034bab418171005b320fbe3344cb9b7845a7373b46a02c483"} Oct 08 14:24:54 crc kubenswrapper[4789]: I1008 14:24:54.141626 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bcb5bfe8-89a1-4085-a486-cce5e4e56789","Type":"ContainerStarted","Data":"3fc6b03bf564dcc0df56c3f69c99dc70645bfef18a6736c3ac35f42547e1cecf"} Oct 08 14:24:54 crc kubenswrapper[4789]: I1008 14:24:54.182267 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.182239491 podStartE2EDuration="2.182239491s" podCreationTimestamp="2025-10-08 14:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:24:54.174774139 +0000 UTC m=+1434.081521631" watchObservedRunningTime="2025-10-08 14:24:54.182239491 +0000 UTC m=+1434.088986983" Oct 08 14:24:54 crc kubenswrapper[4789]: I1008 14:24:54.200282 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.200244479 podStartE2EDuration="2.200244479s" podCreationTimestamp="2025-10-08 14:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:24:54.192041727 +0000 UTC m=+1434.098789219" watchObservedRunningTime="2025-10-08 14:24:54.200244479 +0000 UTC m=+1434.106991991" Oct 08 14:24:57 crc kubenswrapper[4789]: I1008 14:24:57.448107 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 14:24:57 crc kubenswrapper[4789]: I1008 14:24:57.573646 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 14:24:58 crc kubenswrapper[4789]: I1008 14:24:58.443442 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 14:24:58 crc kubenswrapper[4789]: I1008 14:24:58.443491 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 14:24:59 crc kubenswrapper[4789]: I1008 14:24:59.461275 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:24:59 crc kubenswrapper[4789]: I1008 14:24:59.461282 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:25:01 crc kubenswrapper[4789]: I1008 14:25:01.521932 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 14:25:01 crc kubenswrapper[4789]: I1008 14:25:01.522273 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 14:25:01 crc kubenswrapper[4789]: I1008 14:25:01.900357 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-srxvk"] Oct 08 14:25:01 crc kubenswrapper[4789]: I1008 14:25:01.903860 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:01 crc kubenswrapper[4789]: I1008 14:25:01.912491 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-srxvk"] Oct 08 14:25:01 crc kubenswrapper[4789]: I1008 14:25:01.988108 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-utilities\") pod \"community-operators-srxvk\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:01 crc kubenswrapper[4789]: I1008 14:25:01.988370 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trncd\" (UniqueName: \"kubernetes.io/projected/42ef3754-66fe-461d-8956-1f53684bd973-kube-api-access-trncd\") pod \"community-operators-srxvk\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:01 crc kubenswrapper[4789]: I1008 14:25:01.988549 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-catalog-content\") pod \"community-operators-srxvk\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.090330 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trncd\" (UniqueName: \"kubernetes.io/projected/42ef3754-66fe-461d-8956-1f53684bd973-kube-api-access-trncd\") pod \"community-operators-srxvk\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.090648 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-catalog-content\") pod \"community-operators-srxvk\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.090777 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-utilities\") pod \"community-operators-srxvk\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.091145 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-catalog-content\") pod \"community-operators-srxvk\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.092678 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-utilities\") pod \"community-operators-srxvk\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.111905 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trncd\" (UniqueName: \"kubernetes.io/projected/42ef3754-66fe-461d-8956-1f53684bd973-kube-api-access-trncd\") pod \"community-operators-srxvk\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.226143 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.573735 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.591163 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.614158 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.211:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.614256 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.211:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.616179 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 14:25:02 crc kubenswrapper[4789]: W1008 14:25:02.804039 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42ef3754_66fe_461d_8956_1f53684bd973.slice/crio-680f58c41bdbadc81bac589c75ac4dec6f247b2d31f76d51437a7044740aff8b WatchSource:0}: Error finding container 680f58c41bdbadc81bac589c75ac4dec6f247b2d31f76d51437a7044740aff8b: Status 404 returned error can't find the container with id 680f58c41bdbadc81bac589c75ac4dec6f247b2d31f76d51437a7044740aff8b Oct 08 14:25:02 crc kubenswrapper[4789]: I1008 14:25:02.805809 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-srxvk"] Oct 08 14:25:03 crc kubenswrapper[4789]: I1008 14:25:03.229635 4789 generic.go:334] "Generic (PLEG): container finished" podID="42ef3754-66fe-461d-8956-1f53684bd973" containerID="330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5" exitCode=0 Oct 08 14:25:03 crc kubenswrapper[4789]: I1008 14:25:03.229774 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxvk" event={"ID":"42ef3754-66fe-461d-8956-1f53684bd973","Type":"ContainerDied","Data":"330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5"} Oct 08 14:25:03 crc kubenswrapper[4789]: I1008 14:25:03.231077 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxvk" event={"ID":"42ef3754-66fe-461d-8956-1f53684bd973","Type":"ContainerStarted","Data":"680f58c41bdbadc81bac589c75ac4dec6f247b2d31f76d51437a7044740aff8b"} Oct 08 14:25:03 crc kubenswrapper[4789]: I1008 14:25:03.277184 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 14:25:05 crc kubenswrapper[4789]: I1008 14:25:05.252418 4789 generic.go:334] "Generic (PLEG): container finished" podID="42ef3754-66fe-461d-8956-1f53684bd973" containerID="846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd" exitCode=0 Oct 08 14:25:05 crc kubenswrapper[4789]: I1008 14:25:05.252487 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxvk" event={"ID":"42ef3754-66fe-461d-8956-1f53684bd973","Type":"ContainerDied","Data":"846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd"} Oct 08 14:25:06 crc kubenswrapper[4789]: I1008 14:25:06.265255 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxvk" event={"ID":"42ef3754-66fe-461d-8956-1f53684bd973","Type":"ContainerStarted","Data":"adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e"} Oct 08 14:25:06 crc kubenswrapper[4789]: I1008 14:25:06.291416 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-srxvk" podStartSLOduration=2.835932417 podStartE2EDuration="5.291395424s" podCreationTimestamp="2025-10-08 14:25:01 +0000 UTC" firstStartedPulling="2025-10-08 14:25:03.231411776 +0000 UTC m=+1443.138159268" lastFinishedPulling="2025-10-08 14:25:05.686874783 +0000 UTC m=+1445.593622275" observedRunningTime="2025-10-08 14:25:06.280248832 +0000 UTC m=+1446.186996324" watchObservedRunningTime="2025-10-08 14:25:06.291395424 +0000 UTC m=+1446.198142916" Oct 08 14:25:08 crc kubenswrapper[4789]: I1008 14:25:08.447346 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 14:25:08 crc kubenswrapper[4789]: I1008 14:25:08.447698 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 14:25:08 crc kubenswrapper[4789]: I1008 14:25:08.454091 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 14:25:08 crc kubenswrapper[4789]: I1008 14:25:08.454503 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 14:25:10 crc kubenswrapper[4789]: I1008 14:25:10.304828 4789 generic.go:334] "Generic (PLEG): container finished" podID="33326886-bef2-4eb9-903c-79fd1a9f8779" containerID="37f67a89188796a084b2eee0bc252d971771f456ea1e51936cc40b777fd24880" exitCode=137 Oct 08 14:25:10 crc kubenswrapper[4789]: I1008 14:25:10.304905 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"33326886-bef2-4eb9-903c-79fd1a9f8779","Type":"ContainerDied","Data":"37f67a89188796a084b2eee0bc252d971771f456ea1e51936cc40b777fd24880"} Oct 08 14:25:10 crc kubenswrapper[4789]: I1008 14:25:10.882565 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:10 crc kubenswrapper[4789]: I1008 14:25:10.965383 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-config-data\") pod \"33326886-bef2-4eb9-903c-79fd1a9f8779\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " Oct 08 14:25:10 crc kubenswrapper[4789]: I1008 14:25:10.965599 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5sdp\" (UniqueName: \"kubernetes.io/projected/33326886-bef2-4eb9-903c-79fd1a9f8779-kube-api-access-h5sdp\") pod \"33326886-bef2-4eb9-903c-79fd1a9f8779\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " Oct 08 14:25:10 crc kubenswrapper[4789]: I1008 14:25:10.965628 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-combined-ca-bundle\") pod \"33326886-bef2-4eb9-903c-79fd1a9f8779\" (UID: \"33326886-bef2-4eb9-903c-79fd1a9f8779\") " Oct 08 14:25:10 crc kubenswrapper[4789]: I1008 14:25:10.970702 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33326886-bef2-4eb9-903c-79fd1a9f8779-kube-api-access-h5sdp" (OuterVolumeSpecName: "kube-api-access-h5sdp") pod "33326886-bef2-4eb9-903c-79fd1a9f8779" (UID: "33326886-bef2-4eb9-903c-79fd1a9f8779"). InnerVolumeSpecName "kube-api-access-h5sdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:25:10 crc kubenswrapper[4789]: I1008 14:25:10.994092 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-config-data" (OuterVolumeSpecName: "config-data") pod "33326886-bef2-4eb9-903c-79fd1a9f8779" (UID: "33326886-bef2-4eb9-903c-79fd1a9f8779"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:10 crc kubenswrapper[4789]: I1008 14:25:10.995276 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33326886-bef2-4eb9-903c-79fd1a9f8779" (UID: "33326886-bef2-4eb9-903c-79fd1a9f8779"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.068468 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.068504 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5sdp\" (UniqueName: \"kubernetes.io/projected/33326886-bef2-4eb9-903c-79fd1a9f8779-kube-api-access-h5sdp\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.068517 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33326886-bef2-4eb9-903c-79fd1a9f8779-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.319154 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"33326886-bef2-4eb9-903c-79fd1a9f8779","Type":"ContainerDied","Data":"10463ecd989e1c45218ac5336869e47de6292a36c966f9215c27f819fa815ecd"} Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.319258 4789 scope.go:117] "RemoveContainer" containerID="37f67a89188796a084b2eee0bc252d971771f456ea1e51936cc40b777fd24880" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.319279 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.360635 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.380620 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.438322 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 14:25:11 crc kubenswrapper[4789]: E1008 14:25:11.439184 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33326886-bef2-4eb9-903c-79fd1a9f8779" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.439210 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="33326886-bef2-4eb9-903c-79fd1a9f8779" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.439490 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="33326886-bef2-4eb9-903c-79fd1a9f8779" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.440476 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.443654 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.443714 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.443892 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.450014 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.529634 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.530733 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.531571 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.535155 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.582863 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.582910 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.583109 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jks5\" (UniqueName: \"kubernetes.io/projected/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-kube-api-access-6jks5\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.583145 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.583178 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.685122 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jks5\" (UniqueName: \"kubernetes.io/projected/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-kube-api-access-6jks5\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.685165 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.685194 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.685357 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.685376 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.690155 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.692387 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.694903 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.695732 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.704289 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jks5\" (UniqueName: \"kubernetes.io/projected/42136e0e-1d9e-44ea-b9fc-3e7e313fe107-kube-api-access-6jks5\") pod \"nova-cell1-novncproxy-0\" (UID: \"42136e0e-1d9e-44ea-b9fc-3e7e313fe107\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.775205 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:11 crc kubenswrapper[4789]: I1008 14:25:11.853255 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.198:3000/\": dial tcp 10.217.0.198:3000: connect: connection refused" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.227328 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.227386 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.304944 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.366334 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.399779 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.455878 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.496837 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.608736 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-97b9dd99c-x92cg"] Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.610853 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.633947 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-97b9dd99c-x92cg"] Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.661733 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-srxvk"] Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.724457 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-sb\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.724562 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfjsn\" (UniqueName: \"kubernetes.io/projected/4c773033-385f-457a-b976-896f23fb21f8-kube-api-access-nfjsn\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.724780 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-swift-storage-0\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.724836 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-nb\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.724871 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-svc\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.724978 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-config\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.746394 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33326886-bef2-4eb9-903c-79fd1a9f8779" path="/var/lib/kubelet/pods/33326886-bef2-4eb9-903c-79fd1a9f8779/volumes" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.826621 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-sb\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.826741 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfjsn\" (UniqueName: \"kubernetes.io/projected/4c773033-385f-457a-b976-896f23fb21f8-kube-api-access-nfjsn\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.826802 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-swift-storage-0\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.826823 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-nb\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.826842 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-svc\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.826892 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-config\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.829532 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-sb\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.830163 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-nb\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.830208 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-config\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.830425 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-swift-storage-0\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.830435 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-svc\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:12 crc kubenswrapper[4789]: I1008 14:25:12.849104 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfjsn\" (UniqueName: \"kubernetes.io/projected/4c773033-385f-457a-b976-896f23fb21f8-kube-api-access-nfjsn\") pod \"dnsmasq-dns-97b9dd99c-x92cg\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:13 crc kubenswrapper[4789]: I1008 14:25:13.001505 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:13 crc kubenswrapper[4789]: I1008 14:25:13.417411 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"42136e0e-1d9e-44ea-b9fc-3e7e313fe107","Type":"ContainerStarted","Data":"73b858ed00922ede8673241e70c3e4d8330414de6467bfcb35848bd1d8475a5f"} Oct 08 14:25:13 crc kubenswrapper[4789]: I1008 14:25:13.417729 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"42136e0e-1d9e-44ea-b9fc-3e7e313fe107","Type":"ContainerStarted","Data":"bc0b21a50e864b5f8e0293557130cd35a1030867186048ffe3c811e9bb625912"} Oct 08 14:25:13 crc kubenswrapper[4789]: I1008 14:25:13.438943 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.438924112 podStartE2EDuration="2.438924112s" podCreationTimestamp="2025-10-08 14:25:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:25:13.433928357 +0000 UTC m=+1453.340675859" watchObservedRunningTime="2025-10-08 14:25:13.438924112 +0000 UTC m=+1453.345671604" Oct 08 14:25:13 crc kubenswrapper[4789]: I1008 14:25:13.499936 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-97b9dd99c-x92cg"] Oct 08 14:25:14 crc kubenswrapper[4789]: I1008 14:25:14.430312 4789 generic.go:334] "Generic (PLEG): container finished" podID="4c773033-385f-457a-b976-896f23fb21f8" containerID="1cffa181cb7ad075ab5ae7745bdda29a9d6d37be39f2a3b705309451dfe793a4" exitCode=0 Oct 08 14:25:14 crc kubenswrapper[4789]: I1008 14:25:14.430725 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-srxvk" podUID="42ef3754-66fe-461d-8956-1f53684bd973" containerName="registry-server" containerID="cri-o://adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e" gracePeriod=2 Oct 08 14:25:14 crc kubenswrapper[4789]: I1008 14:25:14.430352 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" event={"ID":"4c773033-385f-457a-b976-896f23fb21f8","Type":"ContainerDied","Data":"1cffa181cb7ad075ab5ae7745bdda29a9d6d37be39f2a3b705309451dfe793a4"} Oct 08 14:25:14 crc kubenswrapper[4789]: I1008 14:25:14.431258 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" event={"ID":"4c773033-385f-457a-b976-896f23fb21f8","Type":"ContainerStarted","Data":"b08398dd07b84693f017d0447e4b91769d6a8a7fca50c82c2702210f1b5738ba"} Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.017108 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.043371 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.191550 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-utilities\") pod \"42ef3754-66fe-461d-8956-1f53684bd973\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.191618 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-catalog-content\") pod \"42ef3754-66fe-461d-8956-1f53684bd973\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.191710 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trncd\" (UniqueName: \"kubernetes.io/projected/42ef3754-66fe-461d-8956-1f53684bd973-kube-api-access-trncd\") pod \"42ef3754-66fe-461d-8956-1f53684bd973\" (UID: \"42ef3754-66fe-461d-8956-1f53684bd973\") " Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.192700 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-utilities" (OuterVolumeSpecName: "utilities") pod "42ef3754-66fe-461d-8956-1f53684bd973" (UID: "42ef3754-66fe-461d-8956-1f53684bd973"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.197178 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42ef3754-66fe-461d-8956-1f53684bd973-kube-api-access-trncd" (OuterVolumeSpecName: "kube-api-access-trncd") pod "42ef3754-66fe-461d-8956-1f53684bd973" (UID: "42ef3754-66fe-461d-8956-1f53684bd973"). InnerVolumeSpecName "kube-api-access-trncd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.252636 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42ef3754-66fe-461d-8956-1f53684bd973" (UID: "42ef3754-66fe-461d-8956-1f53684bd973"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.294679 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.294715 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42ef3754-66fe-461d-8956-1f53684bd973-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.294725 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trncd\" (UniqueName: \"kubernetes.io/projected/42ef3754-66fe-461d-8956-1f53684bd973-kube-api-access-trncd\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.442420 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" event={"ID":"4c773033-385f-457a-b976-896f23fb21f8","Type":"ContainerStarted","Data":"c6ef8e9c981ed35929185477cc89625b458f161c50506854c6a9ea680d0d3254"} Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.442591 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.444837 4789 generic.go:334] "Generic (PLEG): container finished" podID="42ef3754-66fe-461d-8956-1f53684bd973" containerID="adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e" exitCode=0 Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.444912 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxvk" event={"ID":"42ef3754-66fe-461d-8956-1f53684bd973","Type":"ContainerDied","Data":"adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e"} Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.444949 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxvk" event={"ID":"42ef3754-66fe-461d-8956-1f53684bd973","Type":"ContainerDied","Data":"680f58c41bdbadc81bac589c75ac4dec6f247b2d31f76d51437a7044740aff8b"} Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.444967 4789 scope.go:117] "RemoveContainer" containerID="adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.445072 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-log" containerID="cri-o://98749808ef5b82555ba4234bb16e54e5ae68c689df50f752330e6b330c7cc4e1" gracePeriod=30 Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.445070 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-api" containerID="cri-o://f4756dad6ac88bdf861036d865cbae522a4d123bb70d7a18a08a318a35635fe3" gracePeriod=30 Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.445120 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srxvk" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.466345 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" podStartSLOduration=3.466328188 podStartE2EDuration="3.466328188s" podCreationTimestamp="2025-10-08 14:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:25:15.464670113 +0000 UTC m=+1455.371417605" watchObservedRunningTime="2025-10-08 14:25:15.466328188 +0000 UTC m=+1455.373075680" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.476955 4789 scope.go:117] "RemoveContainer" containerID="846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.505520 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-srxvk"] Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.509218 4789 scope.go:117] "RemoveContainer" containerID="330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.514757 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-srxvk"] Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.558646 4789 scope.go:117] "RemoveContainer" containerID="adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e" Oct 08 14:25:15 crc kubenswrapper[4789]: E1008 14:25:15.559093 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e\": container with ID starting with adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e not found: ID does not exist" containerID="adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.559132 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e"} err="failed to get container status \"adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e\": rpc error: code = NotFound desc = could not find container \"adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e\": container with ID starting with adf5799198ad6de06af5b13eb2e1ff9b6484240da94e4718260d364f74ddd12e not found: ID does not exist" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.559158 4789 scope.go:117] "RemoveContainer" containerID="846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd" Oct 08 14:25:15 crc kubenswrapper[4789]: E1008 14:25:15.562169 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd\": container with ID starting with 846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd not found: ID does not exist" containerID="846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.562236 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd"} err="failed to get container status \"846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd\": rpc error: code = NotFound desc = could not find container \"846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd\": container with ID starting with 846d3fa7c59cf16e074ef69779d72138b6b51456d85b8ed904a84c35456957cd not found: ID does not exist" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.562265 4789 scope.go:117] "RemoveContainer" containerID="330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5" Oct 08 14:25:15 crc kubenswrapper[4789]: E1008 14:25:15.562560 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5\": container with ID starting with 330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5 not found: ID does not exist" containerID="330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5" Oct 08 14:25:15 crc kubenswrapper[4789]: I1008 14:25:15.562594 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5"} err="failed to get container status \"330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5\": rpc error: code = NotFound desc = could not find container \"330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5\": container with ID starting with 330db5bacfa41847a8125f094902ce4784279a748d908802b9527c4ad7cb7ad5 not found: ID does not exist" Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.455060 4789 generic.go:334] "Generic (PLEG): container finished" podID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerID="f4756dad6ac88bdf861036d865cbae522a4d123bb70d7a18a08a318a35635fe3" exitCode=0 Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.455096 4789 generic.go:334] "Generic (PLEG): container finished" podID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerID="98749808ef5b82555ba4234bb16e54e5ae68c689df50f752330e6b330c7cc4e1" exitCode=143 Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.455145 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb208d86-8b8b-4114-8638-7d8a98a6223f","Type":"ContainerDied","Data":"f4756dad6ac88bdf861036d865cbae522a4d123bb70d7a18a08a318a35635fe3"} Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.455183 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb208d86-8b8b-4114-8638-7d8a98a6223f","Type":"ContainerDied","Data":"98749808ef5b82555ba4234bb16e54e5ae68c689df50f752330e6b330c7cc4e1"} Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.744619 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42ef3754-66fe-461d-8956-1f53684bd973" path="/var/lib/kubelet/pods/42ef3754-66fe-461d-8956-1f53684bd973/volumes" Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.776145 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.839011 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.925714 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-config-data\") pod \"cb208d86-8b8b-4114-8638-7d8a98a6223f\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.925771 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-combined-ca-bundle\") pod \"cb208d86-8b8b-4114-8638-7d8a98a6223f\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.925947 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrdbm\" (UniqueName: \"kubernetes.io/projected/cb208d86-8b8b-4114-8638-7d8a98a6223f-kube-api-access-vrdbm\") pod \"cb208d86-8b8b-4114-8638-7d8a98a6223f\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.926046 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb208d86-8b8b-4114-8638-7d8a98a6223f-logs\") pod \"cb208d86-8b8b-4114-8638-7d8a98a6223f\" (UID: \"cb208d86-8b8b-4114-8638-7d8a98a6223f\") " Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.939757 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb208d86-8b8b-4114-8638-7d8a98a6223f-kube-api-access-vrdbm" (OuterVolumeSpecName: "kube-api-access-vrdbm") pod "cb208d86-8b8b-4114-8638-7d8a98a6223f" (UID: "cb208d86-8b8b-4114-8638-7d8a98a6223f"). InnerVolumeSpecName "kube-api-access-vrdbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.960187 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb208d86-8b8b-4114-8638-7d8a98a6223f-logs" (OuterVolumeSpecName: "logs") pod "cb208d86-8b8b-4114-8638-7d8a98a6223f" (UID: "cb208d86-8b8b-4114-8638-7d8a98a6223f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.982348 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-config-data" (OuterVolumeSpecName: "config-data") pod "cb208d86-8b8b-4114-8638-7d8a98a6223f" (UID: "cb208d86-8b8b-4114-8638-7d8a98a6223f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:16 crc kubenswrapper[4789]: I1008 14:25:16.985760 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb208d86-8b8b-4114-8638-7d8a98a6223f" (UID: "cb208d86-8b8b-4114-8638-7d8a98a6223f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.029184 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.029219 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb208d86-8b8b-4114-8638-7d8a98a6223f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.029256 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrdbm\" (UniqueName: \"kubernetes.io/projected/cb208d86-8b8b-4114-8638-7d8a98a6223f-kube-api-access-vrdbm\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.029265 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb208d86-8b8b-4114-8638-7d8a98a6223f-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.468209 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cb208d86-8b8b-4114-8638-7d8a98a6223f","Type":"ContainerDied","Data":"8ff668821e7a7ad66dab2b17d6c619267cebd0bde0fc9783059f39d2ec5ee224"} Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.468266 4789 scope.go:117] "RemoveContainer" containerID="f4756dad6ac88bdf861036d865cbae522a4d123bb70d7a18a08a318a35635fe3" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.468288 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.494079 4789 scope.go:117] "RemoveContainer" containerID="98749808ef5b82555ba4234bb16e54e5ae68c689df50f752330e6b330c7cc4e1" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.526971 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.547323 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.566668 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:17 crc kubenswrapper[4789]: E1008 14:25:17.567180 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-api" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.567204 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-api" Oct 08 14:25:17 crc kubenswrapper[4789]: E1008 14:25:17.567231 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42ef3754-66fe-461d-8956-1f53684bd973" containerName="extract-content" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.567238 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="42ef3754-66fe-461d-8956-1f53684bd973" containerName="extract-content" Oct 08 14:25:17 crc kubenswrapper[4789]: E1008 14:25:17.567251 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42ef3754-66fe-461d-8956-1f53684bd973" containerName="registry-server" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.567257 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="42ef3754-66fe-461d-8956-1f53684bd973" containerName="registry-server" Oct 08 14:25:17 crc kubenswrapper[4789]: E1008 14:25:17.567273 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42ef3754-66fe-461d-8956-1f53684bd973" containerName="extract-utilities" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.567278 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="42ef3754-66fe-461d-8956-1f53684bd973" containerName="extract-utilities" Oct 08 14:25:17 crc kubenswrapper[4789]: E1008 14:25:17.567298 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-log" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.567303 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-log" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.567489 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-api" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.567504 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="42ef3754-66fe-461d-8956-1f53684bd973" containerName="registry-server" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.567524 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" containerName="nova-api-log" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.568847 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.572845 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.573123 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.573267 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.579037 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.744674 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msrsp\" (UniqueName: \"kubernetes.io/projected/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-kube-api-access-msrsp\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.745143 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.745372 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-config-data\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.745471 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.745576 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-public-tls-certs\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.745712 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-logs\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.847327 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-public-tls-certs\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.847443 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-logs\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.847514 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msrsp\" (UniqueName: \"kubernetes.io/projected/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-kube-api-access-msrsp\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.847606 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.847642 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-config-data\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.847670 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.848314 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-logs\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.852545 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.853013 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-config-data\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.853800 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.863158 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-public-tls-certs\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.872573 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msrsp\" (UniqueName: \"kubernetes.io/projected/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-kube-api-access-msrsp\") pod \"nova-api-0\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " pod="openstack/nova-api-0" Oct 08 14:25:17 crc kubenswrapper[4789]: I1008 14:25:17.884815 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.399822 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:18 crc kubenswrapper[4789]: W1008 14:25:18.475964 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73c1a02c_473c_4bbb_864d_3b07c0fd13e7.slice/crio-84441288fbcec5fded64a50f73b192d4dfbde0d6f290c287c11fd0befda560d3 WatchSource:0}: Error finding container 84441288fbcec5fded64a50f73b192d4dfbde0d6f290c287c11fd0befda560d3: Status 404 returned error can't find the container with id 84441288fbcec5fded64a50f73b192d4dfbde0d6f290c287c11fd0befda560d3 Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.485971 4789 generic.go:334] "Generic (PLEG): container finished" podID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerID="1d0969b22c0457046963c721f2fe13930d401fb05ce2543511f2886ebca4be00" exitCode=137 Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.486033 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerDied","Data":"1d0969b22c0457046963c721f2fe13930d401fb05ce2543511f2886ebca4be00"} Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.654332 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.745408 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb208d86-8b8b-4114-8638-7d8a98a6223f" path="/var/lib/kubelet/pods/cb208d86-8b8b-4114-8638-7d8a98a6223f/volumes" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.786612 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-log-httpd\") pod \"0d62e800-8147-48c2-bab8-6e16c6e22d05\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.786677 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-scripts\") pod \"0d62e800-8147-48c2-bab8-6e16c6e22d05\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.786738 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-run-httpd\") pod \"0d62e800-8147-48c2-bab8-6e16c6e22d05\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.787233 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0d62e800-8147-48c2-bab8-6e16c6e22d05" (UID: "0d62e800-8147-48c2-bab8-6e16c6e22d05"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.787272 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0d62e800-8147-48c2-bab8-6e16c6e22d05" (UID: "0d62e800-8147-48c2-bab8-6e16c6e22d05"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.787356 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-combined-ca-bundle\") pod \"0d62e800-8147-48c2-bab8-6e16c6e22d05\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.787434 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-config-data\") pod \"0d62e800-8147-48c2-bab8-6e16c6e22d05\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.787549 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvcdk\" (UniqueName: \"kubernetes.io/projected/0d62e800-8147-48c2-bab8-6e16c6e22d05-kube-api-access-dvcdk\") pod \"0d62e800-8147-48c2-bab8-6e16c6e22d05\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.787573 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-sg-core-conf-yaml\") pod \"0d62e800-8147-48c2-bab8-6e16c6e22d05\" (UID: \"0d62e800-8147-48c2-bab8-6e16c6e22d05\") " Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.794480 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d62e800-8147-48c2-bab8-6e16c6e22d05-kube-api-access-dvcdk" (OuterVolumeSpecName: "kube-api-access-dvcdk") pod "0d62e800-8147-48c2-bab8-6e16c6e22d05" (UID: "0d62e800-8147-48c2-bab8-6e16c6e22d05"). InnerVolumeSpecName "kube-api-access-dvcdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.794673 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-scripts" (OuterVolumeSpecName: "scripts") pod "0d62e800-8147-48c2-bab8-6e16c6e22d05" (UID: "0d62e800-8147-48c2-bab8-6e16c6e22d05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.797290 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvcdk\" (UniqueName: \"kubernetes.io/projected/0d62e800-8147-48c2-bab8-6e16c6e22d05-kube-api-access-dvcdk\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.797322 4789 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.797333 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.797341 4789 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0d62e800-8147-48c2-bab8-6e16c6e22d05-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.821108 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0d62e800-8147-48c2-bab8-6e16c6e22d05" (UID: "0d62e800-8147-48c2-bab8-6e16c6e22d05"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.892151 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d62e800-8147-48c2-bab8-6e16c6e22d05" (UID: "0d62e800-8147-48c2-bab8-6e16c6e22d05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.898928 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.898959 4789 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:18 crc kubenswrapper[4789]: I1008 14:25:18.929759 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-config-data" (OuterVolumeSpecName: "config-data") pod "0d62e800-8147-48c2-bab8-6e16c6e22d05" (UID: "0d62e800-8147-48c2-bab8-6e16c6e22d05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.002063 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62e800-8147-48c2-bab8-6e16c6e22d05-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.499088 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73c1a02c-473c-4bbb-864d-3b07c0fd13e7","Type":"ContainerStarted","Data":"16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a"} Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.499167 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73c1a02c-473c-4bbb-864d-3b07c0fd13e7","Type":"ContainerStarted","Data":"12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2"} Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.499179 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73c1a02c-473c-4bbb-864d-3b07c0fd13e7","Type":"ContainerStarted","Data":"84441288fbcec5fded64a50f73b192d4dfbde0d6f290c287c11fd0befda560d3"} Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.503445 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0d62e800-8147-48c2-bab8-6e16c6e22d05","Type":"ContainerDied","Data":"82b338c9aa2efcedbe8b04682635b2523691213a211961a37c526c9bad3f239b"} Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.503475 4789 scope.go:117] "RemoveContainer" containerID="461690d543781b3c72717e5cc6694c5a7c839c5e48b71e11ee21c99d3e03b30a" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.503507 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.530786 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.530760843 podStartE2EDuration="2.530760843s" podCreationTimestamp="2025-10-08 14:25:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:25:19.520530195 +0000 UTC m=+1459.427277707" watchObservedRunningTime="2025-10-08 14:25:19.530760843 +0000 UTC m=+1459.437508335" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.548252 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.568900 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.573913 4789 scope.go:117] "RemoveContainer" containerID="0d52a25246184153286f994395ebb0b5b8722d0f40c5743dc54a7c238d1e5f3b" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.580375 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:25:19 crc kubenswrapper[4789]: E1008 14:25:19.581107 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="ceilometer-central-agent" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.581124 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="ceilometer-central-agent" Oct 08 14:25:19 crc kubenswrapper[4789]: E1008 14:25:19.581137 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="sg-core" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.581143 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="sg-core" Oct 08 14:25:19 crc kubenswrapper[4789]: E1008 14:25:19.581176 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="proxy-httpd" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.581182 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="proxy-httpd" Oct 08 14:25:19 crc kubenswrapper[4789]: E1008 14:25:19.581198 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="ceilometer-notification-agent" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.581205 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="ceilometer-notification-agent" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.581382 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="ceilometer-central-agent" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.581398 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="sg-core" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.581408 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="ceilometer-notification-agent" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.581427 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" containerName="proxy-httpd" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.583166 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.585272 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.585479 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.585480 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.592528 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.619870 4789 scope.go:117] "RemoveContainer" containerID="1d0969b22c0457046963c721f2fe13930d401fb05ce2543511f2886ebca4be00" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.645280 4789 scope.go:117] "RemoveContainer" containerID="21ec477643d66c1646db0374733fea8d174901474abbc958fff28425b0b3d3ee" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.717291 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3152634e-3d68-4544-ad93-2999a2273f3a-log-httpd\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.717357 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj2vx\" (UniqueName: \"kubernetes.io/projected/3152634e-3d68-4544-ad93-2999a2273f3a-kube-api-access-fj2vx\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.717391 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.717437 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-config-data\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.717454 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-scripts\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.717524 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3152634e-3d68-4544-ad93-2999a2273f3a-run-httpd\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.717562 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.717612 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.819402 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-config-data\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.819453 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-scripts\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.819478 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3152634e-3d68-4544-ad93-2999a2273f3a-run-httpd\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.819503 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.819553 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.819615 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3152634e-3d68-4544-ad93-2999a2273f3a-log-httpd\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.819645 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj2vx\" (UniqueName: \"kubernetes.io/projected/3152634e-3d68-4544-ad93-2999a2273f3a-kube-api-access-fj2vx\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.819677 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.820365 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3152634e-3d68-4544-ad93-2999a2273f3a-log-httpd\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.820373 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3152634e-3d68-4544-ad93-2999a2273f3a-run-httpd\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.824907 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.825328 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-scripts\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.825506 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.826613 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.826970 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3152634e-3d68-4544-ad93-2999a2273f3a-config-data\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.836512 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj2vx\" (UniqueName: \"kubernetes.io/projected/3152634e-3d68-4544-ad93-2999a2273f3a-kube-api-access-fj2vx\") pod \"ceilometer-0\" (UID: \"3152634e-3d68-4544-ad93-2999a2273f3a\") " pod="openstack/ceilometer-0" Oct 08 14:25:19 crc kubenswrapper[4789]: I1008 14:25:19.903171 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 14:25:20 crc kubenswrapper[4789]: W1008 14:25:20.342047 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3152634e_3d68_4544_ad93_2999a2273f3a.slice/crio-c21b969abe7d99448daf4b7ffa0f0c9ccc365a9ad6548e9bb3b999697e41f3f2 WatchSource:0}: Error finding container c21b969abe7d99448daf4b7ffa0f0c9ccc365a9ad6548e9bb3b999697e41f3f2: Status 404 returned error can't find the container with id c21b969abe7d99448daf4b7ffa0f0c9ccc365a9ad6548e9bb3b999697e41f3f2 Oct 08 14:25:20 crc kubenswrapper[4789]: I1008 14:25:20.343821 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 14:25:20 crc kubenswrapper[4789]: I1008 14:25:20.526777 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3152634e-3d68-4544-ad93-2999a2273f3a","Type":"ContainerStarted","Data":"c21b969abe7d99448daf4b7ffa0f0c9ccc365a9ad6548e9bb3b999697e41f3f2"} Oct 08 14:25:20 crc kubenswrapper[4789]: I1008 14:25:20.765011 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d62e800-8147-48c2-bab8-6e16c6e22d05" path="/var/lib/kubelet/pods/0d62e800-8147-48c2-bab8-6e16c6e22d05/volumes" Oct 08 14:25:21 crc kubenswrapper[4789]: I1008 14:25:21.535376 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3152634e-3d68-4544-ad93-2999a2273f3a","Type":"ContainerStarted","Data":"16e1d744389dd11a022f53d0687ae088e6227bf7eeadbc83ec2445b395b134df"} Oct 08 14:25:21 crc kubenswrapper[4789]: I1008 14:25:21.535675 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3152634e-3d68-4544-ad93-2999a2273f3a","Type":"ContainerStarted","Data":"e73de6e632111d7ae6d262adedeef325b43ee39ac16df2fa248a13ad437febb2"} Oct 08 14:25:21 crc kubenswrapper[4789]: I1008 14:25:21.775901 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:21 crc kubenswrapper[4789]: I1008 14:25:21.795205 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.563594 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.747293 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-4x2xn"] Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.748742 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.753471 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.753642 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.767087 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4x2xn"] Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.894256 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.894561 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-scripts\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.894667 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-config-data\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.894687 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw8sr\" (UniqueName: \"kubernetes.io/projected/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-kube-api-access-zw8sr\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.996909 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.996977 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-scripts\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.997105 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-config-data\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:22 crc kubenswrapper[4789]: I1008 14:25:22.997143 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw8sr\" (UniqueName: \"kubernetes.io/projected/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-kube-api-access-zw8sr\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.003147 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.003601 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-scripts\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.004770 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-config-data\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.016754 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw8sr\" (UniqueName: \"kubernetes.io/projected/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-kube-api-access-zw8sr\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.017296 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-4x2xn\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.076820 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.100840 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65579c8c57-22dsv"] Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.101078 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" podUID="0928b734-4a8d-46df-8726-420d02c65243" containerName="dnsmasq-dns" containerID="cri-o://c752a74e4590f8230f08e6fb24649f87ad81b1253807003ac354a092dfce0b6a" gracePeriod=10 Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.575247 4789 generic.go:334] "Generic (PLEG): container finished" podID="0928b734-4a8d-46df-8726-420d02c65243" containerID="c752a74e4590f8230f08e6fb24649f87ad81b1253807003ac354a092dfce0b6a" exitCode=0 Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.575744 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" event={"ID":"0928b734-4a8d-46df-8726-420d02c65243","Type":"ContainerDied","Data":"c752a74e4590f8230f08e6fb24649f87ad81b1253807003ac354a092dfce0b6a"} Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.598502 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3152634e-3d68-4544-ad93-2999a2273f3a","Type":"ContainerStarted","Data":"868de325868cbeb5e1e9eb2524460c8316a6d320c9709b7d74f1cae811a1b0e8"} Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.599824 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-4x2xn"] Oct 08 14:25:23 crc kubenswrapper[4789]: W1008 14:25:23.607136 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8607d7e0_182c_4dc0_b8a6_2baeb594e61f.slice/crio-09e1d1f700323eceb60dbe0f279b90ff8379a2048e60b7ec313e84d41059eded WatchSource:0}: Error finding container 09e1d1f700323eceb60dbe0f279b90ff8379a2048e60b7ec313e84d41059eded: Status 404 returned error can't find the container with id 09e1d1f700323eceb60dbe0f279b90ff8379a2048e60b7ec313e84d41059eded Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.831087 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.928356 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn67j\" (UniqueName: \"kubernetes.io/projected/0928b734-4a8d-46df-8726-420d02c65243-kube-api-access-sn67j\") pod \"0928b734-4a8d-46df-8726-420d02c65243\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.928482 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-svc\") pod \"0928b734-4a8d-46df-8726-420d02c65243\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.928522 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-swift-storage-0\") pod \"0928b734-4a8d-46df-8726-420d02c65243\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.928568 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-nb\") pod \"0928b734-4a8d-46df-8726-420d02c65243\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.928612 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-config\") pod \"0928b734-4a8d-46df-8726-420d02c65243\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.928730 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-sb\") pod \"0928b734-4a8d-46df-8726-420d02c65243\" (UID: \"0928b734-4a8d-46df-8726-420d02c65243\") " Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.933017 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0928b734-4a8d-46df-8726-420d02c65243-kube-api-access-sn67j" (OuterVolumeSpecName: "kube-api-access-sn67j") pod "0928b734-4a8d-46df-8726-420d02c65243" (UID: "0928b734-4a8d-46df-8726-420d02c65243"). InnerVolumeSpecName "kube-api-access-sn67j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.983075 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-config" (OuterVolumeSpecName: "config") pod "0928b734-4a8d-46df-8726-420d02c65243" (UID: "0928b734-4a8d-46df-8726-420d02c65243"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.983336 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0928b734-4a8d-46df-8726-420d02c65243" (UID: "0928b734-4a8d-46df-8726-420d02c65243"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.983518 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0928b734-4a8d-46df-8726-420d02c65243" (UID: "0928b734-4a8d-46df-8726-420d02c65243"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:25:23 crc kubenswrapper[4789]: I1008 14:25:23.995261 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0928b734-4a8d-46df-8726-420d02c65243" (UID: "0928b734-4a8d-46df-8726-420d02c65243"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.015704 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0928b734-4a8d-46df-8726-420d02c65243" (UID: "0928b734-4a8d-46df-8726-420d02c65243"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.031720 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.031754 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.031765 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.031774 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.031784 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0928b734-4a8d-46df-8726-420d02c65243-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.031795 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn67j\" (UniqueName: \"kubernetes.io/projected/0928b734-4a8d-46df-8726-420d02c65243-kube-api-access-sn67j\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.611171 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3152634e-3d68-4544-ad93-2999a2273f3a","Type":"ContainerStarted","Data":"9bc2272a79a0042d5cb0cce405a60b0f6093ab594bfa809215ea11056c14bfb3"} Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.612641 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.614637 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4x2xn" event={"ID":"8607d7e0-182c-4dc0-b8a6-2baeb594e61f","Type":"ContainerStarted","Data":"62da5d1fa3ec95c2034bc976c16573164d1fbeb0f80e110ec25680f4ae6b2ba1"} Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.614675 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4x2xn" event={"ID":"8607d7e0-182c-4dc0-b8a6-2baeb594e61f","Type":"ContainerStarted","Data":"09e1d1f700323eceb60dbe0f279b90ff8379a2048e60b7ec313e84d41059eded"} Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.618152 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" event={"ID":"0928b734-4a8d-46df-8726-420d02c65243","Type":"ContainerDied","Data":"596f0d3c7bb6bbf347dd7b30e80a116a3094d0ac055075d5bfc1db68d283d1c1"} Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.618185 4789 scope.go:117] "RemoveContainer" containerID="c752a74e4590f8230f08e6fb24649f87ad81b1253807003ac354a092dfce0b6a" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.618292 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65579c8c57-22dsv" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.638332 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.6623069639999999 podStartE2EDuration="5.638307182s" podCreationTimestamp="2025-10-08 14:25:19 +0000 UTC" firstStartedPulling="2025-10-08 14:25:20.345238413 +0000 UTC m=+1460.251985895" lastFinishedPulling="2025-10-08 14:25:24.321238621 +0000 UTC m=+1464.227986113" observedRunningTime="2025-10-08 14:25:24.631732593 +0000 UTC m=+1464.538480085" watchObservedRunningTime="2025-10-08 14:25:24.638307182 +0000 UTC m=+1464.545054694" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.651891 4789 scope.go:117] "RemoveContainer" containerID="8523752a7d14be02c9e4f6b0738d78a955f5224a6e8cfa1d59c45e808998cbcd" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.671403 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-4x2xn" podStartSLOduration=2.671385228 podStartE2EDuration="2.671385228s" podCreationTimestamp="2025-10-08 14:25:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:25:24.652235929 +0000 UTC m=+1464.558983441" watchObservedRunningTime="2025-10-08 14:25:24.671385228 +0000 UTC m=+1464.578132720" Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.682151 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65579c8c57-22dsv"] Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.690265 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65579c8c57-22dsv"] Oct 08 14:25:24 crc kubenswrapper[4789]: I1008 14:25:24.743343 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0928b734-4a8d-46df-8726-420d02c65243" path="/var/lib/kubelet/pods/0928b734-4a8d-46df-8726-420d02c65243/volumes" Oct 08 14:25:27 crc kubenswrapper[4789]: I1008 14:25:27.886241 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 14:25:27 crc kubenswrapper[4789]: I1008 14:25:27.888084 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 14:25:28 crc kubenswrapper[4789]: I1008 14:25:28.918173 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.217:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:25:28 crc kubenswrapper[4789]: I1008 14:25:28.918191 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.217:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:25:29 crc kubenswrapper[4789]: I1008 14:25:29.690566 4789 generic.go:334] "Generic (PLEG): container finished" podID="8607d7e0-182c-4dc0-b8a6-2baeb594e61f" containerID="62da5d1fa3ec95c2034bc976c16573164d1fbeb0f80e110ec25680f4ae6b2ba1" exitCode=0 Oct 08 14:25:29 crc kubenswrapper[4789]: I1008 14:25:29.690677 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4x2xn" event={"ID":"8607d7e0-182c-4dc0-b8a6-2baeb594e61f","Type":"ContainerDied","Data":"62da5d1fa3ec95c2034bc976c16573164d1fbeb0f80e110ec25680f4ae6b2ba1"} Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.082179 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.268766 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw8sr\" (UniqueName: \"kubernetes.io/projected/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-kube-api-access-zw8sr\") pod \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.268844 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-scripts\") pod \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.268959 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-config-data\") pod \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.269115 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-combined-ca-bundle\") pod \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\" (UID: \"8607d7e0-182c-4dc0-b8a6-2baeb594e61f\") " Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.291627 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-scripts" (OuterVolumeSpecName: "scripts") pod "8607d7e0-182c-4dc0-b8a6-2baeb594e61f" (UID: "8607d7e0-182c-4dc0-b8a6-2baeb594e61f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.291707 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-kube-api-access-zw8sr" (OuterVolumeSpecName: "kube-api-access-zw8sr") pod "8607d7e0-182c-4dc0-b8a6-2baeb594e61f" (UID: "8607d7e0-182c-4dc0-b8a6-2baeb594e61f"). InnerVolumeSpecName "kube-api-access-zw8sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.299816 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-config-data" (OuterVolumeSpecName: "config-data") pod "8607d7e0-182c-4dc0-b8a6-2baeb594e61f" (UID: "8607d7e0-182c-4dc0-b8a6-2baeb594e61f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.300369 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8607d7e0-182c-4dc0-b8a6-2baeb594e61f" (UID: "8607d7e0-182c-4dc0-b8a6-2baeb594e61f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.371012 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.371044 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.371057 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw8sr\" (UniqueName: \"kubernetes.io/projected/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-kube-api-access-zw8sr\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.371066 4789 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8607d7e0-182c-4dc0-b8a6-2baeb594e61f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.710388 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-4x2xn" event={"ID":"8607d7e0-182c-4dc0-b8a6-2baeb594e61f","Type":"ContainerDied","Data":"09e1d1f700323eceb60dbe0f279b90ff8379a2048e60b7ec313e84d41059eded"} Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.710425 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09e1d1f700323eceb60dbe0f279b90ff8379a2048e60b7ec313e84d41059eded" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.710438 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-4x2xn" Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.893917 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.894421 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-log" containerID="cri-o://12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2" gracePeriod=30 Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.894470 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-api" containerID="cri-o://16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a" gracePeriod=30 Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.906510 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.906706 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="bcb5bfe8-89a1-4085-a486-cce5e4e56789" containerName="nova-scheduler-scheduler" containerID="cri-o://f34a8ac20d6f6dd034bab418171005b320fbe3344cb9b7845a7373b46a02c483" gracePeriod=30 Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.934485 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.934763 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-log" containerID="cri-o://74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190" gracePeriod=30 Oct 08 14:25:31 crc kubenswrapper[4789]: I1008 14:25:31.934844 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-metadata" containerID="cri-o://c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c" gracePeriod=30 Oct 08 14:25:32 crc kubenswrapper[4789]: E1008 14:25:32.575221 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f34a8ac20d6f6dd034bab418171005b320fbe3344cb9b7845a7373b46a02c483" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 14:25:32 crc kubenswrapper[4789]: E1008 14:25:32.578616 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f34a8ac20d6f6dd034bab418171005b320fbe3344cb9b7845a7373b46a02c483" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 14:25:32 crc kubenswrapper[4789]: E1008 14:25:32.579833 4789 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f34a8ac20d6f6dd034bab418171005b320fbe3344cb9b7845a7373b46a02c483" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 14:25:32 crc kubenswrapper[4789]: E1008 14:25:32.579871 4789 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="bcb5bfe8-89a1-4085-a486-cce5e4e56789" containerName="nova-scheduler-scheduler" Oct 08 14:25:32 crc kubenswrapper[4789]: I1008 14:25:32.720966 4789 generic.go:334] "Generic (PLEG): container finished" podID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerID="12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2" exitCode=143 Oct 08 14:25:32 crc kubenswrapper[4789]: I1008 14:25:32.721026 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73c1a02c-473c-4bbb-864d-3b07c0fd13e7","Type":"ContainerDied","Data":"12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2"} Oct 08 14:25:32 crc kubenswrapper[4789]: I1008 14:25:32.723287 4789 generic.go:334] "Generic (PLEG): container finished" podID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerID="74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190" exitCode=143 Oct 08 14:25:32 crc kubenswrapper[4789]: I1008 14:25:32.723308 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"426d4e55-0509-4936-bede-51b8dc8ccc90","Type":"ContainerDied","Data":"74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190"} Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.270518 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.419296 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-config-data\") pod \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.419368 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msrsp\" (UniqueName: \"kubernetes.io/projected/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-kube-api-access-msrsp\") pod \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.419407 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-public-tls-certs\") pod \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.419445 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-internal-tls-certs\") pod \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.419558 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-combined-ca-bundle\") pod \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.419611 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-logs\") pod \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\" (UID: \"73c1a02c-473c-4bbb-864d-3b07c0fd13e7\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.420754 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-logs" (OuterVolumeSpecName: "logs") pod "73c1a02c-473c-4bbb-864d-3b07c0fd13e7" (UID: "73c1a02c-473c-4bbb-864d-3b07c0fd13e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.425140 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-kube-api-access-msrsp" (OuterVolumeSpecName: "kube-api-access-msrsp") pod "73c1a02c-473c-4bbb-864d-3b07c0fd13e7" (UID: "73c1a02c-473c-4bbb-864d-3b07c0fd13e7"). InnerVolumeSpecName "kube-api-access-msrsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.443887 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": dial tcp 10.217.0.210:8775: connect: connection refused" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.444089 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": dial tcp 10.217.0.210:8775: connect: connection refused" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.451845 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73c1a02c-473c-4bbb-864d-3b07c0fd13e7" (UID: "73c1a02c-473c-4bbb-864d-3b07c0fd13e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.486151 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-config-data" (OuterVolumeSpecName: "config-data") pod "73c1a02c-473c-4bbb-864d-3b07c0fd13e7" (UID: "73c1a02c-473c-4bbb-864d-3b07c0fd13e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.497486 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "73c1a02c-473c-4bbb-864d-3b07c0fd13e7" (UID: "73c1a02c-473c-4bbb-864d-3b07c0fd13e7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.502797 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "73c1a02c-473c-4bbb-864d-3b07c0fd13e7" (UID: "73c1a02c-473c-4bbb-864d-3b07c0fd13e7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.521837 4789 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.521869 4789 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.521881 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.521892 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.521906 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.521918 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msrsp\" (UniqueName: \"kubernetes.io/projected/73c1a02c-473c-4bbb-864d-3b07c0fd13e7-kube-api-access-msrsp\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.665955 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.735787 4789 generic.go:334] "Generic (PLEG): container finished" podID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerID="16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a" exitCode=0 Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.735936 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.738291 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73c1a02c-473c-4bbb-864d-3b07c0fd13e7","Type":"ContainerDied","Data":"16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a"} Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.738357 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"73c1a02c-473c-4bbb-864d-3b07c0fd13e7","Type":"ContainerDied","Data":"84441288fbcec5fded64a50f73b192d4dfbde0d6f290c287c11fd0befda560d3"} Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.738379 4789 scope.go:117] "RemoveContainer" containerID="16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.740762 4789 generic.go:334] "Generic (PLEG): container finished" podID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerID="c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c" exitCode=0 Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.740822 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"426d4e55-0509-4936-bede-51b8dc8ccc90","Type":"ContainerDied","Data":"c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c"} Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.740847 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"426d4e55-0509-4936-bede-51b8dc8ccc90","Type":"ContainerDied","Data":"7f19fe596baf3ca80d468c9d5ca2a7a29668ce99f713f90ae1583ac4d1d2e66c"} Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.740940 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.782738 4789 scope.go:117] "RemoveContainer" containerID="12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.786536 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.803671 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.825215 4789 scope.go:117] "RemoveContainer" containerID="16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.835113 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a\": container with ID starting with 16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a not found: ID does not exist" containerID="16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.835184 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a"} err="failed to get container status \"16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a\": rpc error: code = NotFound desc = could not find container \"16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a\": container with ID starting with 16d77159e6a5584d4d20dbcb03f5a13c7925019f20b35b45b5e5cad061939d0a not found: ID does not exist" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.835213 4789 scope.go:117] "RemoveContainer" containerID="12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.835853 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2\": container with ID starting with 12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2 not found: ID does not exist" containerID="12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.835916 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2"} err="failed to get container status \"12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2\": rpc error: code = NotFound desc = could not find container \"12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2\": container with ID starting with 12e3f8ec9d7962c9c0a197b4851e428dcb8236b987a31bded1fd205dcbd2e1c2 not found: ID does not exist" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.835948 4789 scope.go:117] "RemoveContainer" containerID="c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.838564 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-config-data\") pod \"426d4e55-0509-4936-bede-51b8dc8ccc90\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.838864 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-nova-metadata-tls-certs\") pod \"426d4e55-0509-4936-bede-51b8dc8ccc90\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.838980 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/426d4e55-0509-4936-bede-51b8dc8ccc90-kube-api-access-vnnt5\") pod \"426d4e55-0509-4936-bede-51b8dc8ccc90\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.839074 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/426d4e55-0509-4936-bede-51b8dc8ccc90-logs\") pod \"426d4e55-0509-4936-bede-51b8dc8ccc90\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.839292 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-combined-ca-bundle\") pod \"426d4e55-0509-4936-bede-51b8dc8ccc90\" (UID: \"426d4e55-0509-4936-bede-51b8dc8ccc90\") " Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.848032 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.854495 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-metadata" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.854541 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-metadata" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.854583 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0928b734-4a8d-46df-8726-420d02c65243" containerName="init" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.854593 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0928b734-4a8d-46df-8726-420d02c65243" containerName="init" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.854619 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-log" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.854634 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-log" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.854657 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-api" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.854664 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-api" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.854698 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0928b734-4a8d-46df-8726-420d02c65243" containerName="dnsmasq-dns" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.854707 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0928b734-4a8d-46df-8726-420d02c65243" containerName="dnsmasq-dns" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.854724 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-log" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.854733 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-log" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.854753 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8607d7e0-182c-4dc0-b8a6-2baeb594e61f" containerName="nova-manage" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.854764 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8607d7e0-182c-4dc0-b8a6-2baeb594e61f" containerName="nova-manage" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.855417 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0928b734-4a8d-46df-8726-420d02c65243" containerName="dnsmasq-dns" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.855461 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-log" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.855483 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" containerName="nova-api-api" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.855505 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8607d7e0-182c-4dc0-b8a6-2baeb594e61f" containerName="nova-manage" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.855533 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-metadata" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.855548 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" containerName="nova-metadata-log" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.856804 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/426d4e55-0509-4936-bede-51b8dc8ccc90-logs" (OuterVolumeSpecName: "logs") pod "426d4e55-0509-4936-bede-51b8dc8ccc90" (UID: "426d4e55-0509-4936-bede-51b8dc8ccc90"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.859338 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.863076 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/426d4e55-0509-4936-bede-51b8dc8ccc90-kube-api-access-vnnt5" (OuterVolumeSpecName: "kube-api-access-vnnt5") pod "426d4e55-0509-4936-bede-51b8dc8ccc90" (UID: "426d4e55-0509-4936-bede-51b8dc8ccc90"). InnerVolumeSpecName "kube-api-access-vnnt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.863255 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.863457 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.863458 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.873893 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.885061 4789 scope.go:117] "RemoveContainer" containerID="74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.895666 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-config-data" (OuterVolumeSpecName: "config-data") pod "426d4e55-0509-4936-bede-51b8dc8ccc90" (UID: "426d4e55-0509-4936-bede-51b8dc8ccc90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.904267 4789 scope.go:117] "RemoveContainer" containerID="c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.904685 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c\": container with ID starting with c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c not found: ID does not exist" containerID="c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.904727 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c"} err="failed to get container status \"c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c\": rpc error: code = NotFound desc = could not find container \"c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c\": container with ID starting with c7b1a399f0eb6b0e7488c72082a7d26a6b1e751cb82b15fb07b515cefe17915c not found: ID does not exist" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.904755 4789 scope.go:117] "RemoveContainer" containerID="74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190" Oct 08 14:25:33 crc kubenswrapper[4789]: E1008 14:25:33.905146 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190\": container with ID starting with 74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190 not found: ID does not exist" containerID="74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.905260 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190"} err="failed to get container status \"74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190\": rpc error: code = NotFound desc = could not find container \"74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190\": container with ID starting with 74ec0231a84bfbbb9f5e92a4b61e11a972b55a9e8556052c1b3f655706953190 not found: ID does not exist" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.912898 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "426d4e55-0509-4936-bede-51b8dc8ccc90" (UID: "426d4e55-0509-4936-bede-51b8dc8ccc90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.931727 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "426d4e55-0509-4936-bede-51b8dc8ccc90" (UID: "426d4e55-0509-4936-bede-51b8dc8ccc90"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942249 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-config-data\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942317 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-425lf\" (UniqueName: \"kubernetes.io/projected/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-kube-api-access-425lf\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942370 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942390 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942426 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-logs\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942448 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-public-tls-certs\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942527 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnnt5\" (UniqueName: \"kubernetes.io/projected/426d4e55-0509-4936-bede-51b8dc8ccc90-kube-api-access-vnnt5\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942538 4789 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/426d4e55-0509-4936-bede-51b8dc8ccc90-logs\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942546 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942554 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:33 crc kubenswrapper[4789]: I1008 14:25:33.942561 4789 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/426d4e55-0509-4936-bede-51b8dc8ccc90-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.043375 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.043418 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.043455 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-logs\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.043480 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-public-tls-certs\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.043552 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-config-data\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.043587 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-425lf\" (UniqueName: \"kubernetes.io/projected/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-kube-api-access-425lf\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.044204 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-logs\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.047758 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-public-tls-certs\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.048074 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-config-data\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.048247 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.049499 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.062313 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-425lf\" (UniqueName: \"kubernetes.io/projected/9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c-kube-api-access-425lf\") pod \"nova-api-0\" (UID: \"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c\") " pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.084046 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.092746 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.101714 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.103913 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.107653 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.107659 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.111443 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.145607 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-299t9\" (UniqueName: \"kubernetes.io/projected/9255b8e1-bd22-43cf-b223-87021f091449-kube-api-access-299t9\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.145689 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9255b8e1-bd22-43cf-b223-87021f091449-logs\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.145813 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9255b8e1-bd22-43cf-b223-87021f091449-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.145946 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9255b8e1-bd22-43cf-b223-87021f091449-config-data\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.146028 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9255b8e1-bd22-43cf-b223-87021f091449-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.184079 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.248128 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9255b8e1-bd22-43cf-b223-87021f091449-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.248487 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9255b8e1-bd22-43cf-b223-87021f091449-config-data\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.248524 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9255b8e1-bd22-43cf-b223-87021f091449-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.248611 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-299t9\" (UniqueName: \"kubernetes.io/projected/9255b8e1-bd22-43cf-b223-87021f091449-kube-api-access-299t9\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.248679 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9255b8e1-bd22-43cf-b223-87021f091449-logs\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.249213 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9255b8e1-bd22-43cf-b223-87021f091449-logs\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.256785 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9255b8e1-bd22-43cf-b223-87021f091449-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.256819 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9255b8e1-bd22-43cf-b223-87021f091449-config-data\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.257047 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9255b8e1-bd22-43cf-b223-87021f091449-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.268821 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-299t9\" (UniqueName: \"kubernetes.io/projected/9255b8e1-bd22-43cf-b223-87021f091449-kube-api-access-299t9\") pod \"nova-metadata-0\" (UID: \"9255b8e1-bd22-43cf-b223-87021f091449\") " pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.430816 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.648175 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.746537 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="426d4e55-0509-4936-bede-51b8dc8ccc90" path="/var/lib/kubelet/pods/426d4e55-0509-4936-bede-51b8dc8ccc90/volumes" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.747508 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73c1a02c-473c-4bbb-864d-3b07c0fd13e7" path="/var/lib/kubelet/pods/73c1a02c-473c-4bbb-864d-3b07c0fd13e7/volumes" Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.753809 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c","Type":"ContainerStarted","Data":"816490ab972844c5ef3026a503f981c946c58364197d51ebf84d8dae3af47cd7"} Oct 08 14:25:34 crc kubenswrapper[4789]: I1008 14:25:34.867519 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 14:25:34 crc kubenswrapper[4789]: W1008 14:25:34.873232 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9255b8e1_bd22_43cf_b223_87021f091449.slice/crio-83296117d4c192af842b8bb2229ec7e0612529002aa7f6bdc90a9737c5e15036 WatchSource:0}: Error finding container 83296117d4c192af842b8bb2229ec7e0612529002aa7f6bdc90a9737c5e15036: Status 404 returned error can't find the container with id 83296117d4c192af842b8bb2229ec7e0612529002aa7f6bdc90a9737c5e15036 Oct 08 14:25:35 crc kubenswrapper[4789]: I1008 14:25:35.765201 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9255b8e1-bd22-43cf-b223-87021f091449","Type":"ContainerStarted","Data":"0209b96375b6a0eb0ba1cb6d28b222da0293f70308a8b1e4d96d1971c0da3f1f"} Oct 08 14:25:35 crc kubenswrapper[4789]: I1008 14:25:35.766499 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9255b8e1-bd22-43cf-b223-87021f091449","Type":"ContainerStarted","Data":"15ab37feaec6028a9e18721a2a36ed2e4c96a864cfbbc73324c61f94e801bcea"} Oct 08 14:25:35 crc kubenswrapper[4789]: I1008 14:25:35.766631 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9255b8e1-bd22-43cf-b223-87021f091449","Type":"ContainerStarted","Data":"83296117d4c192af842b8bb2229ec7e0612529002aa7f6bdc90a9737c5e15036"} Oct 08 14:25:35 crc kubenswrapper[4789]: I1008 14:25:35.768189 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c","Type":"ContainerStarted","Data":"c5d954c8a3fac70f8ed1cfe0e47c21406a85fbe68ea8b841eee87749454b5682"} Oct 08 14:25:35 crc kubenswrapper[4789]: I1008 14:25:35.768286 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c","Type":"ContainerStarted","Data":"7b40f552503fb4dc93e0897ee03da945719c4154f87dbd15f204c3da4917e5de"} Oct 08 14:25:35 crc kubenswrapper[4789]: I1008 14:25:35.794750 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.794726378 podStartE2EDuration="1.794726378s" podCreationTimestamp="2025-10-08 14:25:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:25:35.781644623 +0000 UTC m=+1475.688392115" watchObservedRunningTime="2025-10-08 14:25:35.794726378 +0000 UTC m=+1475.701473870" Oct 08 14:25:35 crc kubenswrapper[4789]: I1008 14:25:35.808075 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.808054309 podStartE2EDuration="2.808054309s" podCreationTimestamp="2025-10-08 14:25:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:25:35.803733412 +0000 UTC m=+1475.710480904" watchObservedRunningTime="2025-10-08 14:25:35.808054309 +0000 UTC m=+1475.714801791" Oct 08 14:25:36 crc kubenswrapper[4789]: I1008 14:25:36.779964 4789 generic.go:334] "Generic (PLEG): container finished" podID="bcb5bfe8-89a1-4085-a486-cce5e4e56789" containerID="f34a8ac20d6f6dd034bab418171005b320fbe3344cb9b7845a7373b46a02c483" exitCode=0 Oct 08 14:25:36 crc kubenswrapper[4789]: I1008 14:25:36.780037 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bcb5bfe8-89a1-4085-a486-cce5e4e56789","Type":"ContainerDied","Data":"f34a8ac20d6f6dd034bab418171005b320fbe3344cb9b7845a7373b46a02c483"} Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.401117 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.508866 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw795\" (UniqueName: \"kubernetes.io/projected/bcb5bfe8-89a1-4085-a486-cce5e4e56789-kube-api-access-sw795\") pod \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.508959 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-combined-ca-bundle\") pod \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.509053 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-config-data\") pod \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\" (UID: \"bcb5bfe8-89a1-4085-a486-cce5e4e56789\") " Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.514328 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcb5bfe8-89a1-4085-a486-cce5e4e56789-kube-api-access-sw795" (OuterVolumeSpecName: "kube-api-access-sw795") pod "bcb5bfe8-89a1-4085-a486-cce5e4e56789" (UID: "bcb5bfe8-89a1-4085-a486-cce5e4e56789"). InnerVolumeSpecName "kube-api-access-sw795". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.538085 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-config-data" (OuterVolumeSpecName: "config-data") pod "bcb5bfe8-89a1-4085-a486-cce5e4e56789" (UID: "bcb5bfe8-89a1-4085-a486-cce5e4e56789"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.538673 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcb5bfe8-89a1-4085-a486-cce5e4e56789" (UID: "bcb5bfe8-89a1-4085-a486-cce5e4e56789"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.611290 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw795\" (UniqueName: \"kubernetes.io/projected/bcb5bfe8-89a1-4085-a486-cce5e4e56789-kube-api-access-sw795\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.611327 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.611336 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcb5bfe8-89a1-4085-a486-cce5e4e56789-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.792072 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bcb5bfe8-89a1-4085-a486-cce5e4e56789","Type":"ContainerDied","Data":"3fc6b03bf564dcc0df56c3f69c99dc70645bfef18a6736c3ac35f42547e1cecf"} Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.792130 4789 scope.go:117] "RemoveContainer" containerID="f34a8ac20d6f6dd034bab418171005b320fbe3344cb9b7845a7373b46a02c483" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.792274 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.831851 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.851535 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.869578 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:25:37 crc kubenswrapper[4789]: E1008 14:25:37.870136 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcb5bfe8-89a1-4085-a486-cce5e4e56789" containerName="nova-scheduler-scheduler" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.870156 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcb5bfe8-89a1-4085-a486-cce5e4e56789" containerName="nova-scheduler-scheduler" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.870357 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcb5bfe8-89a1-4085-a486-cce5e4e56789" containerName="nova-scheduler-scheduler" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.871160 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.876652 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 14:25:37 crc kubenswrapper[4789]: I1008 14:25:37.881080 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.019030 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2p2k\" (UniqueName: \"kubernetes.io/projected/ad6bc841-5112-4ce6-a8ae-926e8b7d8043-kube-api-access-s2p2k\") pod \"nova-scheduler-0\" (UID: \"ad6bc841-5112-4ce6-a8ae-926e8b7d8043\") " pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.019503 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6bc841-5112-4ce6-a8ae-926e8b7d8043-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ad6bc841-5112-4ce6-a8ae-926e8b7d8043\") " pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.019548 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad6bc841-5112-4ce6-a8ae-926e8b7d8043-config-data\") pod \"nova-scheduler-0\" (UID: \"ad6bc841-5112-4ce6-a8ae-926e8b7d8043\") " pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.120849 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6bc841-5112-4ce6-a8ae-926e8b7d8043-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ad6bc841-5112-4ce6-a8ae-926e8b7d8043\") " pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.120945 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad6bc841-5112-4ce6-a8ae-926e8b7d8043-config-data\") pod \"nova-scheduler-0\" (UID: \"ad6bc841-5112-4ce6-a8ae-926e8b7d8043\") " pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.121797 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2p2k\" (UniqueName: \"kubernetes.io/projected/ad6bc841-5112-4ce6-a8ae-926e8b7d8043-kube-api-access-s2p2k\") pod \"nova-scheduler-0\" (UID: \"ad6bc841-5112-4ce6-a8ae-926e8b7d8043\") " pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.124611 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad6bc841-5112-4ce6-a8ae-926e8b7d8043-config-data\") pod \"nova-scheduler-0\" (UID: \"ad6bc841-5112-4ce6-a8ae-926e8b7d8043\") " pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.135555 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad6bc841-5112-4ce6-a8ae-926e8b7d8043-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ad6bc841-5112-4ce6-a8ae-926e8b7d8043\") " pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.137791 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2p2k\" (UniqueName: \"kubernetes.io/projected/ad6bc841-5112-4ce6-a8ae-926e8b7d8043-kube-api-access-s2p2k\") pod \"nova-scheduler-0\" (UID: \"ad6bc841-5112-4ce6-a8ae-926e8b7d8043\") " pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.187709 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.638764 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.740616 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcb5bfe8-89a1-4085-a486-cce5e4e56789" path="/var/lib/kubelet/pods/bcb5bfe8-89a1-4085-a486-cce5e4e56789/volumes" Oct 08 14:25:38 crc kubenswrapper[4789]: I1008 14:25:38.802248 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ad6bc841-5112-4ce6-a8ae-926e8b7d8043","Type":"ContainerStarted","Data":"d4aba406630d273ea80a862f6c5c60c1bd1faed70efd1a1abae2ae6628d15d19"} Oct 08 14:25:39 crc kubenswrapper[4789]: I1008 14:25:39.430914 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 14:25:39 crc kubenswrapper[4789]: I1008 14:25:39.431676 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 14:25:39 crc kubenswrapper[4789]: I1008 14:25:39.818156 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ad6bc841-5112-4ce6-a8ae-926e8b7d8043","Type":"ContainerStarted","Data":"19a216e95c6d961aae460ab59004880fe1a2603842c5198f65caac618b5b67b0"} Oct 08 14:25:39 crc kubenswrapper[4789]: I1008 14:25:39.838669 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.838653437 podStartE2EDuration="2.838653437s" podCreationTimestamp="2025-10-08 14:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:25:39.837196437 +0000 UTC m=+1479.743943939" watchObservedRunningTime="2025-10-08 14:25:39.838653437 +0000 UTC m=+1479.745400929" Oct 08 14:25:43 crc kubenswrapper[4789]: I1008 14:25:43.187834 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 14:25:44 crc kubenswrapper[4789]: I1008 14:25:44.184505 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 14:25:44 crc kubenswrapper[4789]: I1008 14:25:44.184827 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 14:25:44 crc kubenswrapper[4789]: I1008 14:25:44.431143 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 14:25:44 crc kubenswrapper[4789]: I1008 14:25:44.431203 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 14:25:45 crc kubenswrapper[4789]: I1008 14:25:45.203242 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:25:45 crc kubenswrapper[4789]: I1008 14:25:45.203294 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:25:45 crc kubenswrapper[4789]: I1008 14:25:45.448197 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9255b8e1-bd22-43cf-b223-87021f091449" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 14:25:45 crc kubenswrapper[4789]: I1008 14:25:45.448236 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9255b8e1-bd22-43cf-b223-87021f091449" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 14:25:48 crc kubenswrapper[4789]: I1008 14:25:48.188343 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 14:25:48 crc kubenswrapper[4789]: I1008 14:25:48.225006 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 14:25:48 crc kubenswrapper[4789]: I1008 14:25:48.929847 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 14:25:49 crc kubenswrapper[4789]: I1008 14:25:49.913780 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.193105 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.193620 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.193847 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.193860 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.203438 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.203781 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.440946 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.443740 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.448671 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 14:25:54 crc kubenswrapper[4789]: I1008 14:25:54.955364 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 14:25:56 crc kubenswrapper[4789]: I1008 14:25:56.433455 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:25:56 crc kubenswrapper[4789]: I1008 14:25:56.433522 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:26:03 crc kubenswrapper[4789]: I1008 14:26:03.682612 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 14:26:05 crc kubenswrapper[4789]: I1008 14:26:05.234737 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 14:26:08 crc kubenswrapper[4789]: I1008 14:26:08.424438 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerName="rabbitmq" containerID="cri-o://2d145f358f5e7b194fe1775e804a9988dab919e072ce986e70c64440037eba38" gracePeriod=604797 Oct 08 14:26:08 crc kubenswrapper[4789]: I1008 14:26:08.798158 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 08 14:26:09 crc kubenswrapper[4789]: I1008 14:26:09.523878 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.110:5671: connect: connection refused" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.132549 4789 generic.go:334] "Generic (PLEG): container finished" podID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerID="2d145f358f5e7b194fe1775e804a9988dab919e072ce986e70c64440037eba38" exitCode=0 Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.132611 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c9336a8d-1117-4da6-a21b-4070f722adc9","Type":"ContainerDied","Data":"2d145f358f5e7b194fe1775e804a9988dab919e072ce986e70c64440037eba38"} Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.132882 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c9336a8d-1117-4da6-a21b-4070f722adc9","Type":"ContainerDied","Data":"f1974ce70aa450a751575356d02ca0d70f7dd4b0cad016edb8124bcf862c705d"} Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.132894 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1974ce70aa450a751575356d02ca0d70f7dd4b0cad016edb8124bcf862c705d" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.173310 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:10 crc kubenswrapper[4789]: E1008 14:26:10.263484 4789 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 08 14:26:10 crc kubenswrapper[4789]: command '/bin/bash -c if [ ! -z "$(cat /etc/pod-info/skipPreStopChecks)" ]; then exit 0; fi; rabbitmq-upgrade await_online_quorum_plus_one -t 604800 && rabbitmq-upgrade await_online_synchronized_mirror -t 604800 || true && rabbitmq-upgrade drain -t 604800' exited with 69: Error: Oct 08 14:26:10 crc kubenswrapper[4789]: {:channel_termination_timeout, {:gen_server, :call, [#PID<13286.4133.0>, {:shutdown, ~c"Node was put into maintenance mode"}, :infinity]}} Oct 08 14:26:10 crc kubenswrapper[4789]: > execCommand=["/bin/bash","-c","if [ ! -z \"$(cat /etc/pod-info/skipPreStopChecks)\" ]; then exit 0; fi; rabbitmq-upgrade await_online_quorum_plus_one -t 604800 \u0026\u0026 rabbitmq-upgrade await_online_synchronized_mirror -t 604800 || true \u0026\u0026 rabbitmq-upgrade drain -t 604800"] containerName="rabbitmq" pod="openstack/rabbitmq-server-0" message=< Oct 08 14:26:10 crc kubenswrapper[4789]: Will wait for a quorum + 1 of nodes to be online for all quorum queues and streams for 604800 seconds... Oct 08 14:26:10 crc kubenswrapper[4789]: Target node seems to be the only one in a single node cluster, the command does not apply Oct 08 14:26:10 crc kubenswrapper[4789]: Will wait for a synchronised mirror be online for all classic mirrored queues for 604800 seconds... Oct 08 14:26:10 crc kubenswrapper[4789]: Target node seems to be the only one in a single node cluster, the command does not apply Oct 08 14:26:10 crc kubenswrapper[4789]: Will put node rabbit@rabbitmq-server-0.rabbitmq-nodes.openstack into maintenance mode. The node will no longer serve any client traffic! Oct 08 14:26:10 crc kubenswrapper[4789]: Error: Oct 08 14:26:10 crc kubenswrapper[4789]: {:channel_termination_timeout, {:gen_server, :call, [#PID<13286.4133.0>, {:shutdown, ~c"Node was put into maintenance mode"}, :infinity]}} Oct 08 14:26:10 crc kubenswrapper[4789]: > Oct 08 14:26:10 crc kubenswrapper[4789]: E1008 14:26:10.263542 4789 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 08 14:26:10 crc kubenswrapper[4789]: command '/bin/bash -c if [ ! -z "$(cat /etc/pod-info/skipPreStopChecks)" ]; then exit 0; fi; rabbitmq-upgrade await_online_quorum_plus_one -t 604800 && rabbitmq-upgrade await_online_synchronized_mirror -t 604800 || true && rabbitmq-upgrade drain -t 604800' exited with 69: Error: Oct 08 14:26:10 crc kubenswrapper[4789]: {:channel_termination_timeout, {:gen_server, :call, [#PID<13286.4133.0>, {:shutdown, ~c"Node was put into maintenance mode"}, :infinity]}} Oct 08 14:26:10 crc kubenswrapper[4789]: > pod="openstack/rabbitmq-server-0" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="rabbitmq" containerID="cri-o://bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.263591 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="rabbitmq" containerID="cri-o://bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b" gracePeriod=604794 Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.343531 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-confd\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.343615 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-plugins\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.343640 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7d6z\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-kube-api-access-k7d6z\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.343667 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-plugins-conf\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.343876 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c9336a8d-1117-4da6-a21b-4070f722adc9-erlang-cookie-secret\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.343908 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-config-data\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.343927 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-tls\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.343962 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-server-conf\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.344042 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-erlang-cookie\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.344085 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c9336a8d-1117-4da6-a21b-4070f722adc9-pod-info\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.344108 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"c9336a8d-1117-4da6-a21b-4070f722adc9\" (UID: \"c9336a8d-1117-4da6-a21b-4070f722adc9\") " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.345825 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.347055 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.348380 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.351765 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.352481 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9336a8d-1117-4da6-a21b-4070f722adc9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.355588 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.355824 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c9336a8d-1117-4da6-a21b-4070f722adc9-pod-info" (OuterVolumeSpecName: "pod-info") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.363040 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-kube-api-access-k7d6z" (OuterVolumeSpecName: "kube-api-access-k7d6z") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "kube-api-access-k7d6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.390085 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-config-data" (OuterVolumeSpecName: "config-data") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.416531 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-server-conf" (OuterVolumeSpecName: "server-conf") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446503 4789 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c9336a8d-1117-4da6-a21b-4070f722adc9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446545 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446554 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446563 4789 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446571 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446581 4789 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c9336a8d-1117-4da6-a21b-4070f722adc9-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446601 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446610 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446619 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7d6z\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-kube-api-access-k7d6z\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.446626 4789 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c9336a8d-1117-4da6-a21b-4070f722adc9-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.479086 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.485273 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c9336a8d-1117-4da6-a21b-4070f722adc9" (UID: "c9336a8d-1117-4da6-a21b-4070f722adc9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.548208 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c9336a8d-1117-4da6-a21b-4070f722adc9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:10 crc kubenswrapper[4789]: I1008 14:26:10.548481 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.141647 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.171494 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.185140 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.195550 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 14:26:11 crc kubenswrapper[4789]: E1008 14:26:11.196338 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerName="rabbitmq" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.196362 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerName="rabbitmq" Oct 08 14:26:11 crc kubenswrapper[4789]: E1008 14:26:11.196426 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerName="setup-container" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.196433 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerName="setup-container" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.196610 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" containerName="rabbitmq" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.197818 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.199816 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.199890 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.199931 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-xrf5t" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.202055 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.202091 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.233056 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.233637 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.234315 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365221 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c4758067-9242-48d2-a066-e007669ef85c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365555 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c4758067-9242-48d2-a066-e007669ef85c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365579 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qqjx\" (UniqueName: \"kubernetes.io/projected/c4758067-9242-48d2-a066-e007669ef85c-kube-api-access-5qqjx\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365662 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365680 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365699 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c4758067-9242-48d2-a066-e007669ef85c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365718 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365829 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365875 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c4758067-9242-48d2-a066-e007669ef85c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.365977 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.366180 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4758067-9242-48d2-a066-e007669ef85c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467539 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c4758067-9242-48d2-a066-e007669ef85c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467580 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c4758067-9242-48d2-a066-e007669ef85c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467599 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qqjx\" (UniqueName: \"kubernetes.io/projected/c4758067-9242-48d2-a066-e007669ef85c-kube-api-access-5qqjx\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467658 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467676 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467699 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c4758067-9242-48d2-a066-e007669ef85c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467721 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467756 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467772 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c4758067-9242-48d2-a066-e007669ef85c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467796 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.467841 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4758067-9242-48d2-a066-e007669ef85c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.468392 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c4758067-9242-48d2-a066-e007669ef85c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.468429 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.468662 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4758067-9242-48d2-a066-e007669ef85c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.468734 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.469509 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c4758067-9242-48d2-a066-e007669ef85c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.469553 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.475576 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.482468 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c4758067-9242-48d2-a066-e007669ef85c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.482824 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c4758067-9242-48d2-a066-e007669ef85c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.485253 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c4758067-9242-48d2-a066-e007669ef85c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.491860 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qqjx\" (UniqueName: \"kubernetes.io/projected/c4758067-9242-48d2-a066-e007669ef85c-kube-api-access-5qqjx\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.576706 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c4758067-9242-48d2-a066-e007669ef85c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:11 crc kubenswrapper[4789]: I1008 14:26:11.849606 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:12 crc kubenswrapper[4789]: I1008 14:26:12.284020 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 14:26:12 crc kubenswrapper[4789]: I1008 14:26:12.742198 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9336a8d-1117-4da6-a21b-4070f722adc9" path="/var/lib/kubelet/pods/c9336a8d-1117-4da6-a21b-4070f722adc9/volumes" Oct 08 14:26:13 crc kubenswrapper[4789]: I1008 14:26:13.161406 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c4758067-9242-48d2-a066-e007669ef85c","Type":"ContainerStarted","Data":"0600fb8659de827572215c86c0567c06320902f3fcd3cb320415a620dab24d83"} Oct 08 14:26:14 crc kubenswrapper[4789]: I1008 14:26:14.172747 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c4758067-9242-48d2-a066-e007669ef85c","Type":"ContainerStarted","Data":"b0b53004c4f49007b0dc4ef121f8fb08580881ef42f40c1e66dfb684f35d90f3"} Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.009760 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069041 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05192759-43d6-4ef0-b6f0-6284e2435317-erlang-cookie-secret\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069145 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-erlang-cookie\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069208 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-plugins-conf\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069233 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtzwg\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-kube-api-access-wtzwg\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069290 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-plugins\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069314 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-confd\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069364 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-server-conf\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069396 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-tls\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069444 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05192759-43d6-4ef0-b6f0-6284e2435317-pod-info\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069494 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.069517 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-config-data\") pod \"05192759-43d6-4ef0-b6f0-6284e2435317\" (UID: \"05192759-43d6-4ef0-b6f0-6284e2435317\") " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.070851 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.076891 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.077060 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.085945 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05192759-43d6-4ef0-b6f0-6284e2435317-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.086426 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.090524 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-kube-api-access-wtzwg" (OuterVolumeSpecName: "kube-api-access-wtzwg") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "kube-api-access-wtzwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.096297 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.102886 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/05192759-43d6-4ef0-b6f0-6284e2435317-pod-info" (OuterVolumeSpecName: "pod-info") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.178508 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.178542 4789 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05192759-43d6-4ef0-b6f0-6284e2435317-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.178568 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.178581 4789 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05192759-43d6-4ef0-b6f0-6284e2435317-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.178595 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.178609 4789 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.178620 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtzwg\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-kube-api-access-wtzwg\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.178634 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.192024 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-config-data" (OuterVolumeSpecName: "config-data") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.205868 4789 generic.go:334] "Generic (PLEG): container finished" podID="05192759-43d6-4ef0-b6f0-6284e2435317" containerID="bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b" exitCode=0 Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.205910 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"05192759-43d6-4ef0-b6f0-6284e2435317","Type":"ContainerDied","Data":"bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b"} Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.205935 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"05192759-43d6-4ef0-b6f0-6284e2435317","Type":"ContainerDied","Data":"35ad962c4a320d3afba8876d91185ee116cfee9f18435274a6efaad6fb89f9c7"} Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.205951 4789 scope.go:117] "RemoveContainer" containerID="bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.206089 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.212621 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-server-conf" (OuterVolumeSpecName: "server-conf") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.213253 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.245214 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "05192759-43d6-4ef0-b6f0-6284e2435317" (UID: "05192759-43d6-4ef0-b6f0-6284e2435317"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.256206 4789 scope.go:117] "RemoveContainer" containerID="257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.280666 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.281187 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.281202 4789 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05192759-43d6-4ef0-b6f0-6284e2435317-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.281214 4789 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05192759-43d6-4ef0-b6f0-6284e2435317-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.303304 4789 scope.go:117] "RemoveContainer" containerID="bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b" Oct 08 14:26:17 crc kubenswrapper[4789]: E1008 14:26:17.303891 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b\": container with ID starting with bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b not found: ID does not exist" containerID="bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.303959 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b"} err="failed to get container status \"bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b\": rpc error: code = NotFound desc = could not find container \"bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b\": container with ID starting with bc861067e4a6860fe13fa60623f2d3008f9fb0e52f073d3e527d6fff7ac7a75b not found: ID does not exist" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.304038 4789 scope.go:117] "RemoveContainer" containerID="257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c" Oct 08 14:26:17 crc kubenswrapper[4789]: E1008 14:26:17.304606 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c\": container with ID starting with 257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c not found: ID does not exist" containerID="257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.304656 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c"} err="failed to get container status \"257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c\": rpc error: code = NotFound desc = could not find container \"257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c\": container with ID starting with 257baeb8db63c053202ca8c96d151264b3aac714975250c408fbe5c3a126661c not found: ID does not exist" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.566336 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.579424 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.594945 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 14:26:17 crc kubenswrapper[4789]: E1008 14:26:17.596084 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="rabbitmq" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.596293 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="rabbitmq" Oct 08 14:26:17 crc kubenswrapper[4789]: E1008 14:26:17.596428 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="setup-container" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.596510 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="setup-container" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.596847 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" containerName="rabbitmq" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.599221 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.606485 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.606699 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.606697 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.606958 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rmgmp" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.606744 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.607018 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.607839 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.614381 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.688700 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.689139 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/698f294c-233d-45d0-a47b-f3580b23b955-pod-info\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.689225 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.689327 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.689495 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bgfg\" (UniqueName: \"kubernetes.io/projected/698f294c-233d-45d0-a47b-f3580b23b955-kube-api-access-2bgfg\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.689616 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/698f294c-233d-45d0-a47b-f3580b23b955-config-data\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.689859 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/698f294c-233d-45d0-a47b-f3580b23b955-server-conf\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.690014 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.690148 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/698f294c-233d-45d0-a47b-f3580b23b955-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.690308 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/698f294c-233d-45d0-a47b-f3580b23b955-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.690465 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.793760 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.793905 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/698f294c-233d-45d0-a47b-f3580b23b955-pod-info\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.793936 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.793963 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.794013 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bgfg\" (UniqueName: \"kubernetes.io/projected/698f294c-233d-45d0-a47b-f3580b23b955-kube-api-access-2bgfg\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.794055 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/698f294c-233d-45d0-a47b-f3580b23b955-config-data\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.794275 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/698f294c-233d-45d0-a47b-f3580b23b955-server-conf\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.794319 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.794360 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/698f294c-233d-45d0-a47b-f3580b23b955-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.794407 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/698f294c-233d-45d0-a47b-f3580b23b955-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.794468 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.794852 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.795235 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.795469 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.797768 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/698f294c-233d-45d0-a47b-f3580b23b955-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.798651 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/698f294c-233d-45d0-a47b-f3580b23b955-config-data\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.798774 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/698f294c-233d-45d0-a47b-f3580b23b955-server-conf\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.801624 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.808845 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/698f294c-233d-45d0-a47b-f3580b23b955-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.808955 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/698f294c-233d-45d0-a47b-f3580b23b955-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.817479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/698f294c-233d-45d0-a47b-f3580b23b955-pod-info\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.818752 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bgfg\" (UniqueName: \"kubernetes.io/projected/698f294c-233d-45d0-a47b-f3580b23b955-kube-api-access-2bgfg\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.868600 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"698f294c-233d-45d0-a47b-f3580b23b955\") " pod="openstack/rabbitmq-server-0" Oct 08 14:26:17 crc kubenswrapper[4789]: I1008 14:26:17.927380 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 14:26:18 crc kubenswrapper[4789]: I1008 14:26:18.401817 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 14:26:18 crc kubenswrapper[4789]: I1008 14:26:18.744152 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05192759-43d6-4ef0-b6f0-6284e2435317" path="/var/lib/kubelet/pods/05192759-43d6-4ef0-b6f0-6284e2435317/volumes" Oct 08 14:26:19 crc kubenswrapper[4789]: I1008 14:26:19.229054 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"698f294c-233d-45d0-a47b-f3580b23b955","Type":"ContainerStarted","Data":"43a5370147124be68fe1034149b291d321fe7f42a3c09adc9406e2ac76df97ff"} Oct 08 14:26:20 crc kubenswrapper[4789]: I1008 14:26:20.245349 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"698f294c-233d-45d0-a47b-f3580b23b955","Type":"ContainerStarted","Data":"01ed87ae3cc1487140c25041eecbdd99b5a1810370c7fd965875a90528b99831"} Oct 08 14:26:20 crc kubenswrapper[4789]: I1008 14:26:20.704578 4789 scope.go:117] "RemoveContainer" containerID="e3f4a50d2a44c2277c7f6e583c4eeef4fadd0de7449e696ac9b577620783eb3f" Oct 08 14:26:20 crc kubenswrapper[4789]: I1008 14:26:20.740831 4789 scope.go:117] "RemoveContainer" containerID="db1238911101afd5fd59caa03415ea3bae71b614c0004c9b526ebc31465a6089" Oct 08 14:26:20 crc kubenswrapper[4789]: I1008 14:26:20.799781 4789 scope.go:117] "RemoveContainer" containerID="2d145f358f5e7b194fe1775e804a9988dab919e072ce986e70c64440037eba38" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.570417 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68c5ffb5c5-mmgjw"] Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.572724 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.580608 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.601356 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68c5ffb5c5-mmgjw"] Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.675054 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-openstack-edpm-ipam\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.675105 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-nb\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.675154 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-sb\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.675178 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-config\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.675279 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-svc\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.675308 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-swift-storage-0\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.675325 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mhqr\" (UniqueName: \"kubernetes.io/projected/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-kube-api-access-7mhqr\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.776924 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-svc\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.777005 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-swift-storage-0\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.777031 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mhqr\" (UniqueName: \"kubernetes.io/projected/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-kube-api-access-7mhqr\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.777076 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-openstack-edpm-ipam\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.777107 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-nb\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.777170 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-sb\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.777199 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-config\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.778424 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-nb\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.778435 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-svc\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.778483 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-sb\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.778501 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-config\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.778618 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-openstack-edpm-ipam\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.778960 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-swift-storage-0\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.812159 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mhqr\" (UniqueName: \"kubernetes.io/projected/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-kube-api-access-7mhqr\") pod \"dnsmasq-dns-68c5ffb5c5-mmgjw\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:21 crc kubenswrapper[4789]: I1008 14:26:21.889705 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:22 crc kubenswrapper[4789]: I1008 14:26:22.374449 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68c5ffb5c5-mmgjw"] Oct 08 14:26:22 crc kubenswrapper[4789]: W1008 14:26:22.387598 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7e4c2ee_ee21_4f9e_abdb_cd74d94d9045.slice/crio-fc354d0a076369c0ceb261e6cf841b7737fe6e0a2204fac407495ec87b8b7455 WatchSource:0}: Error finding container fc354d0a076369c0ceb261e6cf841b7737fe6e0a2204fac407495ec87b8b7455: Status 404 returned error can't find the container with id fc354d0a076369c0ceb261e6cf841b7737fe6e0a2204fac407495ec87b8b7455 Oct 08 14:26:23 crc kubenswrapper[4789]: I1008 14:26:23.273395 4789 generic.go:334] "Generic (PLEG): container finished" podID="a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" containerID="4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5" exitCode=0 Oct 08 14:26:23 crc kubenswrapper[4789]: I1008 14:26:23.273685 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" event={"ID":"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045","Type":"ContainerDied","Data":"4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5"} Oct 08 14:26:23 crc kubenswrapper[4789]: I1008 14:26:23.273715 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" event={"ID":"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045","Type":"ContainerStarted","Data":"fc354d0a076369c0ceb261e6cf841b7737fe6e0a2204fac407495ec87b8b7455"} Oct 08 14:26:24 crc kubenswrapper[4789]: I1008 14:26:24.283888 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" event={"ID":"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045","Type":"ContainerStarted","Data":"d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c"} Oct 08 14:26:24 crc kubenswrapper[4789]: I1008 14:26:24.284341 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:24 crc kubenswrapper[4789]: I1008 14:26:24.312915 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" podStartSLOduration=3.312898392 podStartE2EDuration="3.312898392s" podCreationTimestamp="2025-10-08 14:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:26:24.30471161 +0000 UTC m=+1524.211459112" watchObservedRunningTime="2025-10-08 14:26:24.312898392 +0000 UTC m=+1524.219645884" Oct 08 14:26:26 crc kubenswrapper[4789]: I1008 14:26:26.433080 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:26:26 crc kubenswrapper[4789]: I1008 14:26:26.433443 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:26:31 crc kubenswrapper[4789]: I1008 14:26:31.892192 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:31 crc kubenswrapper[4789]: I1008 14:26:31.964351 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-97b9dd99c-x92cg"] Oct 08 14:26:31 crc kubenswrapper[4789]: I1008 14:26:31.964610 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" podUID="4c773033-385f-457a-b976-896f23fb21f8" containerName="dnsmasq-dns" containerID="cri-o://c6ef8e9c981ed35929185477cc89625b458f161c50506854c6a9ea680d0d3254" gracePeriod=10 Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.096350 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65bf458dd9-l6k8r"] Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.098748 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.121404 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65bf458dd9-l6k8r"] Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.196474 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-ovsdbserver-nb\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.196558 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnkpt\" (UniqueName: \"kubernetes.io/projected/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-kube-api-access-xnkpt\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.196783 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-ovsdbserver-sb\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.196968 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-openstack-edpm-ipam\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.197064 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-dns-svc\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.197167 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-config\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.197360 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-dns-swift-storage-0\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.299152 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-ovsdbserver-sb\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.299553 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-openstack-edpm-ipam\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.299583 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-dns-svc\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.299622 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-config\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.300267 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-ovsdbserver-sb\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.300444 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-dns-svc\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.300458 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-openstack-edpm-ipam\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.300572 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-dns-swift-storage-0\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.300629 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-ovsdbserver-nb\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.300690 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnkpt\" (UniqueName: \"kubernetes.io/projected/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-kube-api-access-xnkpt\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.300805 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-config\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.301328 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-ovsdbserver-nb\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.301345 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-dns-swift-storage-0\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.322284 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnkpt\" (UniqueName: \"kubernetes.io/projected/dbf93231-9ffa-4434-a4c5-3bf97ee66e71-kube-api-access-xnkpt\") pod \"dnsmasq-dns-65bf458dd9-l6k8r\" (UID: \"dbf93231-9ffa-4434-a4c5-3bf97ee66e71\") " pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.361190 4789 generic.go:334] "Generic (PLEG): container finished" podID="4c773033-385f-457a-b976-896f23fb21f8" containerID="c6ef8e9c981ed35929185477cc89625b458f161c50506854c6a9ea680d0d3254" exitCode=0 Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.361243 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" event={"ID":"4c773033-385f-457a-b976-896f23fb21f8","Type":"ContainerDied","Data":"c6ef8e9c981ed35929185477cc89625b458f161c50506854c6a9ea680d0d3254"} Oct 08 14:26:32 crc kubenswrapper[4789]: I1008 14:26:32.471949 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.062476 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65bf458dd9-l6k8r"] Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.313134 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.333920 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfjsn\" (UniqueName: \"kubernetes.io/projected/4c773033-385f-457a-b976-896f23fb21f8-kube-api-access-nfjsn\") pod \"4c773033-385f-457a-b976-896f23fb21f8\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.333979 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-swift-storage-0\") pod \"4c773033-385f-457a-b976-896f23fb21f8\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.334037 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-config\") pod \"4c773033-385f-457a-b976-896f23fb21f8\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.334087 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-svc\") pod \"4c773033-385f-457a-b976-896f23fb21f8\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.334225 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-nb\") pod \"4c773033-385f-457a-b976-896f23fb21f8\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.334251 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-sb\") pod \"4c773033-385f-457a-b976-896f23fb21f8\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.342246 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c773033-385f-457a-b976-896f23fb21f8-kube-api-access-nfjsn" (OuterVolumeSpecName: "kube-api-access-nfjsn") pod "4c773033-385f-457a-b976-896f23fb21f8" (UID: "4c773033-385f-457a-b976-896f23fb21f8"). InnerVolumeSpecName "kube-api-access-nfjsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.392862 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" event={"ID":"dbf93231-9ffa-4434-a4c5-3bf97ee66e71","Type":"ContainerStarted","Data":"cd7bb8031da451c74b77e2bc95c5260d8778ab17ed6cf23a4a193715d8f88e3d"} Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.410076 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" event={"ID":"4c773033-385f-457a-b976-896f23fb21f8","Type":"ContainerDied","Data":"b08398dd07b84693f017d0447e4b91769d6a8a7fca50c82c2702210f1b5738ba"} Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.410311 4789 scope.go:117] "RemoveContainer" containerID="c6ef8e9c981ed35929185477cc89625b458f161c50506854c6a9ea680d0d3254" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.410359 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.412740 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4c773033-385f-457a-b976-896f23fb21f8" (UID: "4c773033-385f-457a-b976-896f23fb21f8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.435424 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-config" (OuterVolumeSpecName: "config") pod "4c773033-385f-457a-b976-896f23fb21f8" (UID: "4c773033-385f-457a-b976-896f23fb21f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.435724 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-config\") pod \"4c773033-385f-457a-b976-896f23fb21f8\" (UID: \"4c773033-385f-457a-b976-896f23fb21f8\") " Oct 08 14:26:33 crc kubenswrapper[4789]: W1008 14:26:33.435853 4789 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/4c773033-385f-457a-b976-896f23fb21f8/volumes/kubernetes.io~configmap/config Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.435871 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-config" (OuterVolumeSpecName: "config") pod "4c773033-385f-457a-b976-896f23fb21f8" (UID: "4c773033-385f-457a-b976-896f23fb21f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.436816 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfjsn\" (UniqueName: \"kubernetes.io/projected/4c773033-385f-457a-b976-896f23fb21f8-kube-api-access-nfjsn\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.436907 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.436979 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.439487 4789 scope.go:117] "RemoveContainer" containerID="1cffa181cb7ad075ab5ae7745bdda29a9d6d37be39f2a3b705309451dfe793a4" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.453901 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4c773033-385f-457a-b976-896f23fb21f8" (UID: "4c773033-385f-457a-b976-896f23fb21f8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.462625 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c773033-385f-457a-b976-896f23fb21f8" (UID: "4c773033-385f-457a-b976-896f23fb21f8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.471690 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4c773033-385f-457a-b976-896f23fb21f8" (UID: "4c773033-385f-457a-b976-896f23fb21f8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.539339 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.539371 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.539381 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c773033-385f-457a-b976-896f23fb21f8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.779501 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-97b9dd99c-x92cg"] Oct 08 14:26:33 crc kubenswrapper[4789]: I1008 14:26:33.788108 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-97b9dd99c-x92cg"] Oct 08 14:26:34 crc kubenswrapper[4789]: I1008 14:26:34.424151 4789 generic.go:334] "Generic (PLEG): container finished" podID="dbf93231-9ffa-4434-a4c5-3bf97ee66e71" containerID="f54a1bca3a1edc274b143f721bf1bbe01adfe092f79153ae9b1c373ae131ab80" exitCode=0 Oct 08 14:26:34 crc kubenswrapper[4789]: I1008 14:26:34.424245 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" event={"ID":"dbf93231-9ffa-4434-a4c5-3bf97ee66e71","Type":"ContainerDied","Data":"f54a1bca3a1edc274b143f721bf1bbe01adfe092f79153ae9b1c373ae131ab80"} Oct 08 14:26:34 crc kubenswrapper[4789]: I1008 14:26:34.742146 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c773033-385f-457a-b976-896f23fb21f8" path="/var/lib/kubelet/pods/4c773033-385f-457a-b976-896f23fb21f8/volumes" Oct 08 14:26:35 crc kubenswrapper[4789]: I1008 14:26:35.433941 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" event={"ID":"dbf93231-9ffa-4434-a4c5-3bf97ee66e71","Type":"ContainerStarted","Data":"45757800f8f1bd4427e13671f2d0b069e9f1d318efeb074bad9d71854569c4a8"} Oct 08 14:26:35 crc kubenswrapper[4789]: I1008 14:26:35.434395 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:35 crc kubenswrapper[4789]: I1008 14:26:35.451555 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" podStartSLOduration=3.451538758 podStartE2EDuration="3.451538758s" podCreationTimestamp="2025-10-08 14:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:26:35.448381593 +0000 UTC m=+1535.355129085" watchObservedRunningTime="2025-10-08 14:26:35.451538758 +0000 UTC m=+1535.358286250" Oct 08 14:26:38 crc kubenswrapper[4789]: I1008 14:26:38.003522 4789 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-97b9dd99c-x92cg" podUID="4c773033-385f-457a-b976-896f23fb21f8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.216:5353: i/o timeout" Oct 08 14:26:42 crc kubenswrapper[4789]: I1008 14:26:42.474637 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65bf458dd9-l6k8r" Oct 08 14:26:42 crc kubenswrapper[4789]: I1008 14:26:42.582542 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68c5ffb5c5-mmgjw"] Oct 08 14:26:42 crc kubenswrapper[4789]: I1008 14:26:42.582841 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" podUID="a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" containerName="dnsmasq-dns" containerID="cri-o://d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c" gracePeriod=10 Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.149934 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.323963 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-sb\") pod \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.324096 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-config\") pod \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.324146 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-swift-storage-0\") pod \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.324169 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-nb\") pod \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.324254 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-openstack-edpm-ipam\") pod \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.324322 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mhqr\" (UniqueName: \"kubernetes.io/projected/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-kube-api-access-7mhqr\") pod \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.324349 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-svc\") pod \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\" (UID: \"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045\") " Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.333249 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-kube-api-access-7mhqr" (OuterVolumeSpecName: "kube-api-access-7mhqr") pod "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" (UID: "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045"). InnerVolumeSpecName "kube-api-access-7mhqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.375969 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" (UID: "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.376581 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" (UID: "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.381094 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-config" (OuterVolumeSpecName: "config") pod "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" (UID: "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.382343 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" (UID: "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.383390 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" (UID: "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.384883 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" (UID: "a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.427829 4789 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.427869 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.427889 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.427902 4789 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.427915 4789 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.427928 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.427945 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mhqr\" (UniqueName: \"kubernetes.io/projected/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045-kube-api-access-7mhqr\") on node \"crc\" DevicePath \"\"" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.574961 4789 generic.go:334] "Generic (PLEG): container finished" podID="a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" containerID="d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c" exitCode=0 Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.575027 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.575047 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" event={"ID":"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045","Type":"ContainerDied","Data":"d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c"} Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.575079 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c5ffb5c5-mmgjw" event={"ID":"a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045","Type":"ContainerDied","Data":"fc354d0a076369c0ceb261e6cf841b7737fe6e0a2204fac407495ec87b8b7455"} Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.575112 4789 scope.go:117] "RemoveContainer" containerID="d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.599129 4789 scope.go:117] "RemoveContainer" containerID="4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.609830 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68c5ffb5c5-mmgjw"] Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.619536 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68c5ffb5c5-mmgjw"] Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.634282 4789 scope.go:117] "RemoveContainer" containerID="d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c" Oct 08 14:26:43 crc kubenswrapper[4789]: E1008 14:26:43.634743 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c\": container with ID starting with d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c not found: ID does not exist" containerID="d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.634776 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c"} err="failed to get container status \"d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c\": rpc error: code = NotFound desc = could not find container \"d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c\": container with ID starting with d227bd39cbdbd16436ceb82fa194300638bc7a79a999699f8e7d3f33310c3a7c not found: ID does not exist" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.634796 4789 scope.go:117] "RemoveContainer" containerID="4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5" Oct 08 14:26:43 crc kubenswrapper[4789]: E1008 14:26:43.635100 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5\": container with ID starting with 4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5 not found: ID does not exist" containerID="4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5" Oct 08 14:26:43 crc kubenswrapper[4789]: I1008 14:26:43.635126 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5"} err="failed to get container status \"4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5\": rpc error: code = NotFound desc = could not find container \"4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5\": container with ID starting with 4c09deb28e184c45bd5eef4242b9b665c950ca5c0ef56d9c8730c25c037d0fc5 not found: ID does not exist" Oct 08 14:26:44 crc kubenswrapper[4789]: I1008 14:26:44.740759 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" path="/var/lib/kubelet/pods/a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045/volumes" Oct 08 14:26:46 crc kubenswrapper[4789]: I1008 14:26:46.618725 4789 generic.go:334] "Generic (PLEG): container finished" podID="c4758067-9242-48d2-a066-e007669ef85c" containerID="b0b53004c4f49007b0dc4ef121f8fb08580881ef42f40c1e66dfb684f35d90f3" exitCode=0 Oct 08 14:26:46 crc kubenswrapper[4789]: I1008 14:26:46.618771 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c4758067-9242-48d2-a066-e007669ef85c","Type":"ContainerDied","Data":"b0b53004c4f49007b0dc4ef121f8fb08580881ef42f40c1e66dfb684f35d90f3"} Oct 08 14:26:47 crc kubenswrapper[4789]: I1008 14:26:47.630277 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c4758067-9242-48d2-a066-e007669ef85c","Type":"ContainerStarted","Data":"91fed3f97d975b0a7e8633bef45359e78eaaccee9fe401ea6b91d481a1e4d4e6"} Oct 08 14:26:47 crc kubenswrapper[4789]: I1008 14:26:47.631169 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:26:47 crc kubenswrapper[4789]: I1008 14:26:47.659376 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.659352655 podStartE2EDuration="36.659352655s" podCreationTimestamp="2025-10-08 14:26:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:26:47.650610588 +0000 UTC m=+1547.557358080" watchObservedRunningTime="2025-10-08 14:26:47.659352655 +0000 UTC m=+1547.566100147" Oct 08 14:26:52 crc kubenswrapper[4789]: I1008 14:26:52.680663 4789 generic.go:334] "Generic (PLEG): container finished" podID="698f294c-233d-45d0-a47b-f3580b23b955" containerID="01ed87ae3cc1487140c25041eecbdd99b5a1810370c7fd965875a90528b99831" exitCode=0 Oct 08 14:26:52 crc kubenswrapper[4789]: I1008 14:26:52.680760 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"698f294c-233d-45d0-a47b-f3580b23b955","Type":"ContainerDied","Data":"01ed87ae3cc1487140c25041eecbdd99b5a1810370c7fd965875a90528b99831"} Oct 08 14:26:53 crc kubenswrapper[4789]: I1008 14:26:53.696217 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"698f294c-233d-45d0-a47b-f3580b23b955","Type":"ContainerStarted","Data":"6998789156a699c5483e8208f7a6cabcfb81980dbb328c66a4a664e4f37cfc69"} Oct 08 14:26:53 crc kubenswrapper[4789]: I1008 14:26:53.696894 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 14:26:53 crc kubenswrapper[4789]: I1008 14:26:53.719364 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.719343333 podStartE2EDuration="36.719343333s" podCreationTimestamp="2025-10-08 14:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:26:53.71919873 +0000 UTC m=+1553.625946222" watchObservedRunningTime="2025-10-08 14:26:53.719343333 +0000 UTC m=+1553.626090835" Oct 08 14:26:56 crc kubenswrapper[4789]: I1008 14:26:56.432967 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:26:56 crc kubenswrapper[4789]: I1008 14:26:56.433331 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:26:56 crc kubenswrapper[4789]: I1008 14:26:56.433396 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:26:56 crc kubenswrapper[4789]: I1008 14:26:56.434368 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:26:56 crc kubenswrapper[4789]: I1008 14:26:56.434450 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" gracePeriod=600 Oct 08 14:26:56 crc kubenswrapper[4789]: E1008 14:26:56.596045 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:26:56 crc kubenswrapper[4789]: I1008 14:26:56.728684 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" exitCode=0 Oct 08 14:26:56 crc kubenswrapper[4789]: I1008 14:26:56.728747 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30"} Oct 08 14:26:56 crc kubenswrapper[4789]: I1008 14:26:56.729116 4789 scope.go:117] "RemoveContainer" containerID="26322f12c01ca82f351a4ae80e71651fb887b0762ae10b5b893ab0870c97be77" Oct 08 14:26:56 crc kubenswrapper[4789]: I1008 14:26:56.729665 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:26:56 crc kubenswrapper[4789]: E1008 14:26:56.729965 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.864060 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr"] Oct 08 14:27:00 crc kubenswrapper[4789]: E1008 14:27:00.865251 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c773033-385f-457a-b976-896f23fb21f8" containerName="dnsmasq-dns" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.865269 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c773033-385f-457a-b976-896f23fb21f8" containerName="dnsmasq-dns" Oct 08 14:27:00 crc kubenswrapper[4789]: E1008 14:27:00.865293 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c773033-385f-457a-b976-896f23fb21f8" containerName="init" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.865299 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c773033-385f-457a-b976-896f23fb21f8" containerName="init" Oct 08 14:27:00 crc kubenswrapper[4789]: E1008 14:27:00.865334 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" containerName="dnsmasq-dns" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.865341 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" containerName="dnsmasq-dns" Oct 08 14:27:00 crc kubenswrapper[4789]: E1008 14:27:00.865352 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" containerName="init" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.865359 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" containerName="init" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.865598 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c773033-385f-457a-b976-896f23fb21f8" containerName="dnsmasq-dns" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.865613 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e4c2ee-ee21-4f9e-abdb-cd74d94d9045" containerName="dnsmasq-dns" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.866435 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.879187 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.881320 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.881347 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.881627 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.894312 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr"] Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.987528 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.987608 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-489kx\" (UniqueName: \"kubernetes.io/projected/484052d0-6aeb-4f53-bad2-d6228aaadcf3-kube-api-access-489kx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.988069 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:00 crc kubenswrapper[4789]: I1008 14:27:00.988238 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.089788 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.090145 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.090177 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-489kx\" (UniqueName: \"kubernetes.io/projected/484052d0-6aeb-4f53-bad2-d6228aaadcf3-kube-api-access-489kx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.090254 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.099073 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.099341 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.109861 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-489kx\" (UniqueName: \"kubernetes.io/projected/484052d0-6aeb-4f53-bad2-d6228aaadcf3-kube-api-access-489kx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.110011 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.186100 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:01 crc kubenswrapper[4789]: W1008 14:27:01.784026 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod484052d0_6aeb_4f53_bad2_d6228aaadcf3.slice/crio-06af64661cf60ffb2a27374c8a57a53e29470cfad72a646138eccd78ed14ecd4 WatchSource:0}: Error finding container 06af64661cf60ffb2a27374c8a57a53e29470cfad72a646138eccd78ed14ecd4: Status 404 returned error can't find the container with id 06af64661cf60ffb2a27374c8a57a53e29470cfad72a646138eccd78ed14ecd4 Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.784676 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr"] Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.786273 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 14:27:01 crc kubenswrapper[4789]: I1008 14:27:01.853172 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 14:27:02 crc kubenswrapper[4789]: I1008 14:27:02.793429 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" event={"ID":"484052d0-6aeb-4f53-bad2-d6228aaadcf3","Type":"ContainerStarted","Data":"06af64661cf60ffb2a27374c8a57a53e29470cfad72a646138eccd78ed14ecd4"} Oct 08 14:27:07 crc kubenswrapper[4789]: I1008 14:27:07.933256 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 14:27:08 crc kubenswrapper[4789]: I1008 14:27:08.730490 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:27:08 crc kubenswrapper[4789]: E1008 14:27:08.730760 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:27:15 crc kubenswrapper[4789]: E1008 14:27:15.644656 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Oct 08 14:27:15 crc kubenswrapper[4789]: E1008 14:27:15.645301 4789 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 08 14:27:15 crc kubenswrapper[4789]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Oct 08 14:27:15 crc kubenswrapper[4789]: - hosts: all Oct 08 14:27:15 crc kubenswrapper[4789]: strategy: linear Oct 08 14:27:15 crc kubenswrapper[4789]: tasks: Oct 08 14:27:15 crc kubenswrapper[4789]: - name: Enable podified-repos Oct 08 14:27:15 crc kubenswrapper[4789]: become: true Oct 08 14:27:15 crc kubenswrapper[4789]: ansible.builtin.shell: | Oct 08 14:27:15 crc kubenswrapper[4789]: set -euxo pipefail Oct 08 14:27:15 crc kubenswrapper[4789]: pushd /var/tmp Oct 08 14:27:15 crc kubenswrapper[4789]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Oct 08 14:27:15 crc kubenswrapper[4789]: pushd repo-setup-main Oct 08 14:27:15 crc kubenswrapper[4789]: python3 -m venv ./venv Oct 08 14:27:15 crc kubenswrapper[4789]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Oct 08 14:27:15 crc kubenswrapper[4789]: ./venv/bin/repo-setup current-podified -b antelope Oct 08 14:27:15 crc kubenswrapper[4789]: popd Oct 08 14:27:15 crc kubenswrapper[4789]: rm -rf repo-setup-main Oct 08 14:27:15 crc kubenswrapper[4789]: Oct 08 14:27:15 crc kubenswrapper[4789]: Oct 08 14:27:15 crc kubenswrapper[4789]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Oct 08 14:27:15 crc kubenswrapper[4789]: edpm_override_hosts: openstack-edpm-ipam Oct 08 14:27:15 crc kubenswrapper[4789]: edpm_service_type: repo-setup Oct 08 14:27:15 crc kubenswrapper[4789]: Oct 08 14:27:15 crc kubenswrapper[4789]: Oct 08 14:27:15 crc kubenswrapper[4789]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-489kx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr_openstack(484052d0-6aeb-4f53-bad2-d6228aaadcf3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Oct 08 14:27:15 crc kubenswrapper[4789]: > logger="UnhandledError" Oct 08 14:27:15 crc kubenswrapper[4789]: E1008 14:27:15.646460 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" podUID="484052d0-6aeb-4f53-bad2-d6228aaadcf3" Oct 08 14:27:15 crc kubenswrapper[4789]: E1008 14:27:15.955933 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" podUID="484052d0-6aeb-4f53-bad2-d6228aaadcf3" Oct 08 14:27:19 crc kubenswrapper[4789]: I1008 14:27:19.730378 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:27:19 crc kubenswrapper[4789]: E1008 14:27:19.731000 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:27:20 crc kubenswrapper[4789]: I1008 14:27:20.960298 4789 scope.go:117] "RemoveContainer" containerID="914304771c99361e295955a77713afedc158bd52c56496f402ef0de96bc15cf1" Oct 08 14:27:31 crc kubenswrapper[4789]: I1008 14:27:31.097586 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" event={"ID":"484052d0-6aeb-4f53-bad2-d6228aaadcf3","Type":"ContainerStarted","Data":"6897dd4ee6b38c236c1af2cccaef5f57a89b74ecc35c4b0ba6f55e08394cbab9"} Oct 08 14:27:31 crc kubenswrapper[4789]: I1008 14:27:31.123713 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" podStartSLOduration=2.7207588449999998 podStartE2EDuration="31.123694362s" podCreationTimestamp="2025-10-08 14:27:00 +0000 UTC" firstStartedPulling="2025-10-08 14:27:01.786087508 +0000 UTC m=+1561.692835000" lastFinishedPulling="2025-10-08 14:27:30.189023025 +0000 UTC m=+1590.095770517" observedRunningTime="2025-10-08 14:27:31.114056021 +0000 UTC m=+1591.020803513" watchObservedRunningTime="2025-10-08 14:27:31.123694362 +0000 UTC m=+1591.030441854" Oct 08 14:27:34 crc kubenswrapper[4789]: I1008 14:27:34.730432 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:27:34 crc kubenswrapper[4789]: E1008 14:27:34.731251 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:27:42 crc kubenswrapper[4789]: I1008 14:27:42.202845 4789 generic.go:334] "Generic (PLEG): container finished" podID="484052d0-6aeb-4f53-bad2-d6228aaadcf3" containerID="6897dd4ee6b38c236c1af2cccaef5f57a89b74ecc35c4b0ba6f55e08394cbab9" exitCode=0 Oct 08 14:27:42 crc kubenswrapper[4789]: I1008 14:27:42.202951 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" event={"ID":"484052d0-6aeb-4f53-bad2-d6228aaadcf3","Type":"ContainerDied","Data":"6897dd4ee6b38c236c1af2cccaef5f57a89b74ecc35c4b0ba6f55e08394cbab9"} Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.705225 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.741104 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-ssh-key\") pod \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.741163 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-repo-setup-combined-ca-bundle\") pod \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.741202 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-inventory\") pod \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.741293 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-489kx\" (UniqueName: \"kubernetes.io/projected/484052d0-6aeb-4f53-bad2-d6228aaadcf3-kube-api-access-489kx\") pod \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\" (UID: \"484052d0-6aeb-4f53-bad2-d6228aaadcf3\") " Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.746855 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "484052d0-6aeb-4f53-bad2-d6228aaadcf3" (UID: "484052d0-6aeb-4f53-bad2-d6228aaadcf3"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.746946 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/484052d0-6aeb-4f53-bad2-d6228aaadcf3-kube-api-access-489kx" (OuterVolumeSpecName: "kube-api-access-489kx") pod "484052d0-6aeb-4f53-bad2-d6228aaadcf3" (UID: "484052d0-6aeb-4f53-bad2-d6228aaadcf3"). InnerVolumeSpecName "kube-api-access-489kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.770193 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "484052d0-6aeb-4f53-bad2-d6228aaadcf3" (UID: "484052d0-6aeb-4f53-bad2-d6228aaadcf3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.773669 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-inventory" (OuterVolumeSpecName: "inventory") pod "484052d0-6aeb-4f53-bad2-d6228aaadcf3" (UID: "484052d0-6aeb-4f53-bad2-d6228aaadcf3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.843742 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.843775 4789 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.843787 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/484052d0-6aeb-4f53-bad2-d6228aaadcf3-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:27:43 crc kubenswrapper[4789]: I1008 14:27:43.843799 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-489kx\" (UniqueName: \"kubernetes.io/projected/484052d0-6aeb-4f53-bad2-d6228aaadcf3-kube-api-access-489kx\") on node \"crc\" DevicePath \"\"" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.221776 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" event={"ID":"484052d0-6aeb-4f53-bad2-d6228aaadcf3","Type":"ContainerDied","Data":"06af64661cf60ffb2a27374c8a57a53e29470cfad72a646138eccd78ed14ecd4"} Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.221817 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06af64661cf60ffb2a27374c8a57a53e29470cfad72a646138eccd78ed14ecd4" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.221864 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.312961 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj"] Oct 08 14:27:44 crc kubenswrapper[4789]: E1008 14:27:44.313494 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="484052d0-6aeb-4f53-bad2-d6228aaadcf3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.313517 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="484052d0-6aeb-4f53-bad2-d6228aaadcf3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.313838 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="484052d0-6aeb-4f53-bad2-d6228aaadcf3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.314730 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.318601 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.318601 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.318914 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.318948 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.323364 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj"] Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.354265 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-785jj\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.354391 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhb4l\" (UniqueName: \"kubernetes.io/projected/c5a284ba-b502-45cc-b9b6-f460908bee2f-kube-api-access-xhb4l\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-785jj\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.354449 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-785jj\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.455448 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-785jj\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.455528 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhb4l\" (UniqueName: \"kubernetes.io/projected/c5a284ba-b502-45cc-b9b6-f460908bee2f-kube-api-access-xhb4l\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-785jj\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.455566 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-785jj\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.459444 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-785jj\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.459485 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-785jj\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.479190 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhb4l\" (UniqueName: \"kubernetes.io/projected/c5a284ba-b502-45cc-b9b6-f460908bee2f-kube-api-access-xhb4l\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-785jj\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:44 crc kubenswrapper[4789]: I1008 14:27:44.630220 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:45 crc kubenswrapper[4789]: I1008 14:27:45.175282 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj"] Oct 08 14:27:45 crc kubenswrapper[4789]: W1008 14:27:45.178098 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5a284ba_b502_45cc_b9b6_f460908bee2f.slice/crio-7bac7be58b735cc0dbcc5b355ddb350ecf24e424ac626e11f5c1065c363992ee WatchSource:0}: Error finding container 7bac7be58b735cc0dbcc5b355ddb350ecf24e424ac626e11f5c1065c363992ee: Status 404 returned error can't find the container with id 7bac7be58b735cc0dbcc5b355ddb350ecf24e424ac626e11f5c1065c363992ee Oct 08 14:27:45 crc kubenswrapper[4789]: I1008 14:27:45.232115 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" event={"ID":"c5a284ba-b502-45cc-b9b6-f460908bee2f","Type":"ContainerStarted","Data":"7bac7be58b735cc0dbcc5b355ddb350ecf24e424ac626e11f5c1065c363992ee"} Oct 08 14:27:46 crc kubenswrapper[4789]: I1008 14:27:46.244519 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" event={"ID":"c5a284ba-b502-45cc-b9b6-f460908bee2f","Type":"ContainerStarted","Data":"2b17636fb5659e1b4ef50d1129fa5b964517a922222fef755cce043cf4674e53"} Oct 08 14:27:46 crc kubenswrapper[4789]: I1008 14:27:46.265556 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" podStartSLOduration=1.804284686 podStartE2EDuration="2.265530534s" podCreationTimestamp="2025-10-08 14:27:44 +0000 UTC" firstStartedPulling="2025-10-08 14:27:45.18077413 +0000 UTC m=+1605.087521622" lastFinishedPulling="2025-10-08 14:27:45.642019958 +0000 UTC m=+1605.548767470" observedRunningTime="2025-10-08 14:27:46.261599007 +0000 UTC m=+1606.168346509" watchObservedRunningTime="2025-10-08 14:27:46.265530534 +0000 UTC m=+1606.172278036" Oct 08 14:27:46 crc kubenswrapper[4789]: I1008 14:27:46.731115 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:27:46 crc kubenswrapper[4789]: E1008 14:27:46.731564 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:27:49 crc kubenswrapper[4789]: I1008 14:27:49.272288 4789 generic.go:334] "Generic (PLEG): container finished" podID="c5a284ba-b502-45cc-b9b6-f460908bee2f" containerID="2b17636fb5659e1b4ef50d1129fa5b964517a922222fef755cce043cf4674e53" exitCode=0 Oct 08 14:27:49 crc kubenswrapper[4789]: I1008 14:27:49.272374 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" event={"ID":"c5a284ba-b502-45cc-b9b6-f460908bee2f","Type":"ContainerDied","Data":"2b17636fb5659e1b4ef50d1129fa5b964517a922222fef755cce043cf4674e53"} Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.707529 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.871814 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-inventory\") pod \"c5a284ba-b502-45cc-b9b6-f460908bee2f\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.872009 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhb4l\" (UniqueName: \"kubernetes.io/projected/c5a284ba-b502-45cc-b9b6-f460908bee2f-kube-api-access-xhb4l\") pod \"c5a284ba-b502-45cc-b9b6-f460908bee2f\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.872113 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-ssh-key\") pod \"c5a284ba-b502-45cc-b9b6-f460908bee2f\" (UID: \"c5a284ba-b502-45cc-b9b6-f460908bee2f\") " Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.886329 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5a284ba-b502-45cc-b9b6-f460908bee2f-kube-api-access-xhb4l" (OuterVolumeSpecName: "kube-api-access-xhb4l") pod "c5a284ba-b502-45cc-b9b6-f460908bee2f" (UID: "c5a284ba-b502-45cc-b9b6-f460908bee2f"). InnerVolumeSpecName "kube-api-access-xhb4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.908041 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-inventory" (OuterVolumeSpecName: "inventory") pod "c5a284ba-b502-45cc-b9b6-f460908bee2f" (UID: "c5a284ba-b502-45cc-b9b6-f460908bee2f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.909854 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c5a284ba-b502-45cc-b9b6-f460908bee2f" (UID: "c5a284ba-b502-45cc-b9b6-f460908bee2f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.977241 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.977293 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhb4l\" (UniqueName: \"kubernetes.io/projected/c5a284ba-b502-45cc-b9b6-f460908bee2f-kube-api-access-xhb4l\") on node \"crc\" DevicePath \"\"" Oct 08 14:27:50 crc kubenswrapper[4789]: I1008 14:27:50.977313 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5a284ba-b502-45cc-b9b6-f460908bee2f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.298521 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" event={"ID":"c5a284ba-b502-45cc-b9b6-f460908bee2f","Type":"ContainerDied","Data":"7bac7be58b735cc0dbcc5b355ddb350ecf24e424ac626e11f5c1065c363992ee"} Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.298568 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bac7be58b735cc0dbcc5b355ddb350ecf24e424ac626e11f5c1065c363992ee" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.298579 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-785jj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.380680 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj"] Oct 08 14:27:51 crc kubenswrapper[4789]: E1008 14:27:51.381158 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a284ba-b502-45cc-b9b6-f460908bee2f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.381176 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a284ba-b502-45cc-b9b6-f460908bee2f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.381387 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a284ba-b502-45cc-b9b6-f460908bee2f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.382091 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.384673 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.384793 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.385445 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.386285 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.395738 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj"] Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.486966 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2shjp\" (UniqueName: \"kubernetes.io/projected/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-kube-api-access-2shjp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.487775 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.488069 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.488392 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.590516 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.590610 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.590710 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2shjp\" (UniqueName: \"kubernetes.io/projected/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-kube-api-access-2shjp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.590791 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.596695 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.596807 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.597279 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.607223 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2shjp\" (UniqueName: \"kubernetes.io/projected/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-kube-api-access-2shjp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:51 crc kubenswrapper[4789]: I1008 14:27:51.708667 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:27:52 crc kubenswrapper[4789]: I1008 14:27:52.230933 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj"] Oct 08 14:27:52 crc kubenswrapper[4789]: I1008 14:27:52.316174 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" event={"ID":"5f34622a-a071-45a1-b7fd-75fa8fbb70fe","Type":"ContainerStarted","Data":"1a56b1f3a0c82b78fae88d73abedef4849e9c59f4608915fac2854d1332d4afa"} Oct 08 14:27:53 crc kubenswrapper[4789]: I1008 14:27:53.330127 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" event={"ID":"5f34622a-a071-45a1-b7fd-75fa8fbb70fe","Type":"ContainerStarted","Data":"f4a6021200876b90bc644aab857483db581ee64d37314bbbd56442cc9cfb8bf4"} Oct 08 14:27:53 crc kubenswrapper[4789]: I1008 14:27:53.351927 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" podStartSLOduration=1.902093767 podStartE2EDuration="2.351906985s" podCreationTimestamp="2025-10-08 14:27:51 +0000 UTC" firstStartedPulling="2025-10-08 14:27:52.23353004 +0000 UTC m=+1612.140277532" lastFinishedPulling="2025-10-08 14:27:52.683343258 +0000 UTC m=+1612.590090750" observedRunningTime="2025-10-08 14:27:53.348010689 +0000 UTC m=+1613.254758201" watchObservedRunningTime="2025-10-08 14:27:53.351906985 +0000 UTC m=+1613.258654497" Oct 08 14:28:01 crc kubenswrapper[4789]: I1008 14:28:01.730192 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:28:01 crc kubenswrapper[4789]: E1008 14:28:01.733834 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:28:13 crc kubenswrapper[4789]: I1008 14:28:13.731178 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:28:13 crc kubenswrapper[4789]: E1008 14:28:13.732143 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.691201 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-crwd4"] Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.694519 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.708974 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-crwd4"] Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.819186 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-catalog-content\") pod \"redhat-operators-crwd4\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.819289 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-utilities\") pod \"redhat-operators-crwd4\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.819365 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45w8h\" (UniqueName: \"kubernetes.io/projected/83e76522-68cc-43e7-9811-7ed7441fefc3-kube-api-access-45w8h\") pod \"redhat-operators-crwd4\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.884899 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8vfkq"] Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.887107 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.897718 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vfkq"] Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.921178 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45w8h\" (UniqueName: \"kubernetes.io/projected/83e76522-68cc-43e7-9811-7ed7441fefc3-kube-api-access-45w8h\") pod \"redhat-operators-crwd4\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.921290 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-catalog-content\") pod \"redhat-operators-crwd4\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.921349 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-utilities\") pod \"redhat-operators-crwd4\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.921838 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-utilities\") pod \"redhat-operators-crwd4\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.922091 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-catalog-content\") pod \"redhat-operators-crwd4\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:16 crc kubenswrapper[4789]: I1008 14:28:16.947482 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45w8h\" (UniqueName: \"kubernetes.io/projected/83e76522-68cc-43e7-9811-7ed7441fefc3-kube-api-access-45w8h\") pod \"redhat-operators-crwd4\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.021813 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.023034 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-catalog-content\") pod \"certified-operators-8vfkq\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.023101 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-utilities\") pod \"certified-operators-8vfkq\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.023218 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwzgs\" (UniqueName: \"kubernetes.io/projected/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-kube-api-access-kwzgs\") pod \"certified-operators-8vfkq\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.125700 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwzgs\" (UniqueName: \"kubernetes.io/projected/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-kube-api-access-kwzgs\") pod \"certified-operators-8vfkq\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.125859 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-catalog-content\") pod \"certified-operators-8vfkq\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.125935 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-utilities\") pod \"certified-operators-8vfkq\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.127161 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-utilities\") pod \"certified-operators-8vfkq\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.127229 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-catalog-content\") pod \"certified-operators-8vfkq\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.165422 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwzgs\" (UniqueName: \"kubernetes.io/projected/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-kube-api-access-kwzgs\") pod \"certified-operators-8vfkq\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.208132 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.526967 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vfkq"] Oct 08 14:28:17 crc kubenswrapper[4789]: W1008 14:28:17.577618 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83e76522_68cc_43e7_9811_7ed7441fefc3.slice/crio-afa0367c4d00a85d9730d3f61f3c1ea48aef0d696c26b854410dca5ffac9008a WatchSource:0}: Error finding container afa0367c4d00a85d9730d3f61f3c1ea48aef0d696c26b854410dca5ffac9008a: Status 404 returned error can't find the container with id afa0367c4d00a85d9730d3f61f3c1ea48aef0d696c26b854410dca5ffac9008a Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.578735 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-crwd4"] Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.596448 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vfkq" event={"ID":"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2","Type":"ContainerStarted","Data":"6c463b60a3f435cc06463bfc3170dbee1959563a5f075adc1d417f8d7c76b717"} Oct 08 14:28:17 crc kubenswrapper[4789]: I1008 14:28:17.598320 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crwd4" event={"ID":"83e76522-68cc-43e7-9811-7ed7441fefc3","Type":"ContainerStarted","Data":"afa0367c4d00a85d9730d3f61f3c1ea48aef0d696c26b854410dca5ffac9008a"} Oct 08 14:28:18 crc kubenswrapper[4789]: I1008 14:28:18.612266 4789 generic.go:334] "Generic (PLEG): container finished" podID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerID="246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6" exitCode=0 Oct 08 14:28:18 crc kubenswrapper[4789]: I1008 14:28:18.612331 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crwd4" event={"ID":"83e76522-68cc-43e7-9811-7ed7441fefc3","Type":"ContainerDied","Data":"246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6"} Oct 08 14:28:18 crc kubenswrapper[4789]: I1008 14:28:18.614283 4789 generic.go:334] "Generic (PLEG): container finished" podID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerID="ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388" exitCode=0 Oct 08 14:28:18 crc kubenswrapper[4789]: I1008 14:28:18.614324 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vfkq" event={"ID":"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2","Type":"ContainerDied","Data":"ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388"} Oct 08 14:28:20 crc kubenswrapper[4789]: I1008 14:28:20.650198 4789 generic.go:334] "Generic (PLEG): container finished" podID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerID="7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e" exitCode=0 Oct 08 14:28:20 crc kubenswrapper[4789]: I1008 14:28:20.651290 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vfkq" event={"ID":"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2","Type":"ContainerDied","Data":"7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e"} Oct 08 14:28:20 crc kubenswrapper[4789]: I1008 14:28:20.673239 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crwd4" event={"ID":"83e76522-68cc-43e7-9811-7ed7441fefc3","Type":"ContainerStarted","Data":"26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d"} Oct 08 14:28:21 crc kubenswrapper[4789]: I1008 14:28:21.696759 4789 generic.go:334] "Generic (PLEG): container finished" podID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerID="26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d" exitCode=0 Oct 08 14:28:21 crc kubenswrapper[4789]: I1008 14:28:21.696825 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crwd4" event={"ID":"83e76522-68cc-43e7-9811-7ed7441fefc3","Type":"ContainerDied","Data":"26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d"} Oct 08 14:28:21 crc kubenswrapper[4789]: I1008 14:28:21.699341 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vfkq" event={"ID":"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2","Type":"ContainerStarted","Data":"98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e"} Oct 08 14:28:21 crc kubenswrapper[4789]: I1008 14:28:21.753322 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8vfkq" podStartSLOduration=3.047192656 podStartE2EDuration="5.75330752s" podCreationTimestamp="2025-10-08 14:28:16 +0000 UTC" firstStartedPulling="2025-10-08 14:28:18.616949456 +0000 UTC m=+1638.523696938" lastFinishedPulling="2025-10-08 14:28:21.32306431 +0000 UTC m=+1641.229811802" observedRunningTime="2025-10-08 14:28:21.751404018 +0000 UTC m=+1641.658151540" watchObservedRunningTime="2025-10-08 14:28:21.75330752 +0000 UTC m=+1641.660055012" Oct 08 14:28:22 crc kubenswrapper[4789]: I1008 14:28:22.710185 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crwd4" event={"ID":"83e76522-68cc-43e7-9811-7ed7441fefc3","Type":"ContainerStarted","Data":"a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26"} Oct 08 14:28:22 crc kubenswrapper[4789]: I1008 14:28:22.731833 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-crwd4" podStartSLOduration=3.211413904 podStartE2EDuration="6.731819285s" podCreationTimestamp="2025-10-08 14:28:16 +0000 UTC" firstStartedPulling="2025-10-08 14:28:18.620291057 +0000 UTC m=+1638.527038539" lastFinishedPulling="2025-10-08 14:28:22.140696428 +0000 UTC m=+1642.047443920" observedRunningTime="2025-10-08 14:28:22.726364076 +0000 UTC m=+1642.633111568" watchObservedRunningTime="2025-10-08 14:28:22.731819285 +0000 UTC m=+1642.638566767" Oct 08 14:28:26 crc kubenswrapper[4789]: I1008 14:28:26.730024 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:28:26 crc kubenswrapper[4789]: E1008 14:28:26.730782 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:28:27 crc kubenswrapper[4789]: I1008 14:28:27.022598 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:27 crc kubenswrapper[4789]: I1008 14:28:27.022670 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:27 crc kubenswrapper[4789]: I1008 14:28:27.209033 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:27 crc kubenswrapper[4789]: I1008 14:28:27.209569 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:27 crc kubenswrapper[4789]: I1008 14:28:27.252384 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:27 crc kubenswrapper[4789]: I1008 14:28:27.807286 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:27 crc kubenswrapper[4789]: I1008 14:28:27.866521 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vfkq"] Oct 08 14:28:28 crc kubenswrapper[4789]: I1008 14:28:28.069274 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-crwd4" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerName="registry-server" probeResult="failure" output=< Oct 08 14:28:28 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 14:28:28 crc kubenswrapper[4789]: > Oct 08 14:28:29 crc kubenswrapper[4789]: I1008 14:28:29.775186 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8vfkq" podUID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerName="registry-server" containerID="cri-o://98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e" gracePeriod=2 Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.206047 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.341151 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-utilities\") pod \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.341379 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-catalog-content\") pod \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.341467 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwzgs\" (UniqueName: \"kubernetes.io/projected/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-kube-api-access-kwzgs\") pod \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\" (UID: \"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2\") " Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.342231 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-utilities" (OuterVolumeSpecName: "utilities") pod "6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" (UID: "6e9bb868-6cb7-475d-93c6-2e4306eb7ee2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.346776 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-kube-api-access-kwzgs" (OuterVolumeSpecName: "kube-api-access-kwzgs") pod "6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" (UID: "6e9bb868-6cb7-475d-93c6-2e4306eb7ee2"). InnerVolumeSpecName "kube-api-access-kwzgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.390121 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" (UID: "6e9bb868-6cb7-475d-93c6-2e4306eb7ee2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.443859 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.443904 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwzgs\" (UniqueName: \"kubernetes.io/projected/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-kube-api-access-kwzgs\") on node \"crc\" DevicePath \"\"" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.443919 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.788208 4789 generic.go:334] "Generic (PLEG): container finished" podID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerID="98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e" exitCode=0 Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.788271 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vfkq" event={"ID":"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2","Type":"ContainerDied","Data":"98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e"} Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.788311 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vfkq" event={"ID":"6e9bb868-6cb7-475d-93c6-2e4306eb7ee2","Type":"ContainerDied","Data":"6c463b60a3f435cc06463bfc3170dbee1959563a5f075adc1d417f8d7c76b717"} Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.788336 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vfkq" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.788338 4789 scope.go:117] "RemoveContainer" containerID="98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.828042 4789 scope.go:117] "RemoveContainer" containerID="7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.829646 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vfkq"] Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.837854 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8vfkq"] Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.849253 4789 scope.go:117] "RemoveContainer" containerID="ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.900581 4789 scope.go:117] "RemoveContainer" containerID="98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e" Oct 08 14:28:30 crc kubenswrapper[4789]: E1008 14:28:30.901912 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e\": container with ID starting with 98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e not found: ID does not exist" containerID="98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.901947 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e"} err="failed to get container status \"98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e\": rpc error: code = NotFound desc = could not find container \"98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e\": container with ID starting with 98f14c638066b3d5cc07a2c0ce1937ee292b84aae6ceeb5f896f73cbce6f8a5e not found: ID does not exist" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.901969 4789 scope.go:117] "RemoveContainer" containerID="7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e" Oct 08 14:28:30 crc kubenswrapper[4789]: E1008 14:28:30.902373 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e\": container with ID starting with 7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e not found: ID does not exist" containerID="7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.902402 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e"} err="failed to get container status \"7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e\": rpc error: code = NotFound desc = could not find container \"7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e\": container with ID starting with 7df8d4ec8480474e3a93393b0d5330e6070d723bcc16cb92419e1d41b915f83e not found: ID does not exist" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.902417 4789 scope.go:117] "RemoveContainer" containerID="ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388" Oct 08 14:28:30 crc kubenswrapper[4789]: E1008 14:28:30.902677 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388\": container with ID starting with ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388 not found: ID does not exist" containerID="ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388" Oct 08 14:28:30 crc kubenswrapper[4789]: I1008 14:28:30.902703 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388"} err="failed to get container status \"ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388\": rpc error: code = NotFound desc = could not find container \"ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388\": container with ID starting with ce050ee86e46dfd9ecda26b581809cbd2f7547f93bdbe5622b50308c2da6a388 not found: ID does not exist" Oct 08 14:28:32 crc kubenswrapper[4789]: I1008 14:28:32.749208 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" path="/var/lib/kubelet/pods/6e9bb868-6cb7-475d-93c6-2e4306eb7ee2/volumes" Oct 08 14:28:37 crc kubenswrapper[4789]: I1008 14:28:37.077935 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:37 crc kubenswrapper[4789]: I1008 14:28:37.146633 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:37 crc kubenswrapper[4789]: I1008 14:28:37.316464 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-crwd4"] Oct 08 14:28:37 crc kubenswrapper[4789]: I1008 14:28:37.730219 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:28:37 crc kubenswrapper[4789]: E1008 14:28:37.730572 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:28:38 crc kubenswrapper[4789]: I1008 14:28:38.879832 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-crwd4" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerName="registry-server" containerID="cri-o://a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26" gracePeriod=2 Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.330330 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.516807 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-utilities\") pod \"83e76522-68cc-43e7-9811-7ed7441fefc3\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.516868 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45w8h\" (UniqueName: \"kubernetes.io/projected/83e76522-68cc-43e7-9811-7ed7441fefc3-kube-api-access-45w8h\") pod \"83e76522-68cc-43e7-9811-7ed7441fefc3\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.516968 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-catalog-content\") pod \"83e76522-68cc-43e7-9811-7ed7441fefc3\" (UID: \"83e76522-68cc-43e7-9811-7ed7441fefc3\") " Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.517505 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-utilities" (OuterVolumeSpecName: "utilities") pod "83e76522-68cc-43e7-9811-7ed7441fefc3" (UID: "83e76522-68cc-43e7-9811-7ed7441fefc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.522837 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83e76522-68cc-43e7-9811-7ed7441fefc3-kube-api-access-45w8h" (OuterVolumeSpecName: "kube-api-access-45w8h") pod "83e76522-68cc-43e7-9811-7ed7441fefc3" (UID: "83e76522-68cc-43e7-9811-7ed7441fefc3"). InnerVolumeSpecName "kube-api-access-45w8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.592293 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83e76522-68cc-43e7-9811-7ed7441fefc3" (UID: "83e76522-68cc-43e7-9811-7ed7441fefc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.618627 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.618665 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45w8h\" (UniqueName: \"kubernetes.io/projected/83e76522-68cc-43e7-9811-7ed7441fefc3-kube-api-access-45w8h\") on node \"crc\" DevicePath \"\"" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.618677 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e76522-68cc-43e7-9811-7ed7441fefc3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.889890 4789 generic.go:334] "Generic (PLEG): container finished" podID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerID="a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26" exitCode=0 Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.889933 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crwd4" event={"ID":"83e76522-68cc-43e7-9811-7ed7441fefc3","Type":"ContainerDied","Data":"a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26"} Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.889969 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-crwd4" event={"ID":"83e76522-68cc-43e7-9811-7ed7441fefc3","Type":"ContainerDied","Data":"afa0367c4d00a85d9730d3f61f3c1ea48aef0d696c26b854410dca5ffac9008a"} Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.890005 4789 scope.go:117] "RemoveContainer" containerID="a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.890029 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-crwd4" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.927053 4789 scope.go:117] "RemoveContainer" containerID="26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.937380 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-crwd4"] Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.947175 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-crwd4"] Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.953001 4789 scope.go:117] "RemoveContainer" containerID="246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.993855 4789 scope.go:117] "RemoveContainer" containerID="a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26" Oct 08 14:28:39 crc kubenswrapper[4789]: E1008 14:28:39.995553 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26\": container with ID starting with a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26 not found: ID does not exist" containerID="a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.995592 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26"} err="failed to get container status \"a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26\": rpc error: code = NotFound desc = could not find container \"a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26\": container with ID starting with a1ebde6a3fab4df7039a0de849c474cbe602f17032e93ff6330b198ac4b02a26 not found: ID does not exist" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.995619 4789 scope.go:117] "RemoveContainer" containerID="26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d" Oct 08 14:28:39 crc kubenswrapper[4789]: E1008 14:28:39.996093 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d\": container with ID starting with 26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d not found: ID does not exist" containerID="26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.996135 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d"} err="failed to get container status \"26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d\": rpc error: code = NotFound desc = could not find container \"26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d\": container with ID starting with 26c2f20e8cf9b64a1b9c5d90ad2c0d77013e33198c9f53989823c7029f76a05d not found: ID does not exist" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.996161 4789 scope.go:117] "RemoveContainer" containerID="246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6" Oct 08 14:28:39 crc kubenswrapper[4789]: E1008 14:28:39.996421 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6\": container with ID starting with 246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6 not found: ID does not exist" containerID="246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6" Oct 08 14:28:39 crc kubenswrapper[4789]: I1008 14:28:39.996445 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6"} err="failed to get container status \"246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6\": rpc error: code = NotFound desc = could not find container \"246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6\": container with ID starting with 246357843e526e5ea9bc02a423160b38f5973f2003e8503525152e6dc8d797a6 not found: ID does not exist" Oct 08 14:28:40 crc kubenswrapper[4789]: I1008 14:28:40.744799 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" path="/var/lib/kubelet/pods/83e76522-68cc-43e7-9811-7ed7441fefc3/volumes" Oct 08 14:28:52 crc kubenswrapper[4789]: I1008 14:28:52.730755 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:28:52 crc kubenswrapper[4789]: E1008 14:28:52.731721 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:29:07 crc kubenswrapper[4789]: I1008 14:29:07.729975 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:29:07 crc kubenswrapper[4789]: E1008 14:29:07.730907 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:29:22 crc kubenswrapper[4789]: I1008 14:29:22.733194 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:29:22 crc kubenswrapper[4789]: E1008 14:29:22.733938 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:29:27 crc kubenswrapper[4789]: I1008 14:29:27.037048 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-kxpvf"] Oct 08 14:29:27 crc kubenswrapper[4789]: I1008 14:29:27.046874 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-kxpvf"] Oct 08 14:29:28 crc kubenswrapper[4789]: I1008 14:29:28.742504 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8f44a9a-deb5-478b-8414-f103280bdf3f" path="/var/lib/kubelet/pods/b8f44a9a-deb5-478b-8414-f103280bdf3f/volumes" Oct 08 14:29:29 crc kubenswrapper[4789]: I1008 14:29:29.027492 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-f6hnn"] Oct 08 14:29:29 crc kubenswrapper[4789]: I1008 14:29:29.038399 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-f6hnn"] Oct 08 14:29:30 crc kubenswrapper[4789]: I1008 14:29:30.743798 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14b540ee-ce54-430d-8313-d7b33bcdb1a7" path="/var/lib/kubelet/pods/14b540ee-ce54-430d-8313-d7b33bcdb1a7/volumes" Oct 08 14:29:34 crc kubenswrapper[4789]: I1008 14:29:34.730433 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:29:34 crc kubenswrapper[4789]: E1008 14:29:34.731262 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:29:43 crc kubenswrapper[4789]: I1008 14:29:43.030426 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-pd9bc"] Oct 08 14:29:43 crc kubenswrapper[4789]: I1008 14:29:43.046935 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-2d7c-account-create-cgqlh"] Oct 08 14:29:43 crc kubenswrapper[4789]: I1008 14:29:43.056582 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-7de8-account-create-4zc9n"] Oct 08 14:29:43 crc kubenswrapper[4789]: I1008 14:29:43.064737 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-pd9bc"] Oct 08 14:29:43 crc kubenswrapper[4789]: I1008 14:29:43.073606 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-7de8-account-create-4zc9n"] Oct 08 14:29:43 crc kubenswrapper[4789]: I1008 14:29:43.081954 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-2d7c-account-create-cgqlh"] Oct 08 14:29:44 crc kubenswrapper[4789]: I1008 14:29:44.741496 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdab5d2d-97d5-4c68-ad3e-2589adfcadd4" path="/var/lib/kubelet/pods/bdab5d2d-97d5-4c68-ad3e-2589adfcadd4/volumes" Oct 08 14:29:44 crc kubenswrapper[4789]: I1008 14:29:44.744008 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f0c4cf-3ebe-440d-a47a-ad8966d5087e" path="/var/lib/kubelet/pods/d3f0c4cf-3ebe-440d-a47a-ad8966d5087e/volumes" Oct 08 14:29:44 crc kubenswrapper[4789]: I1008 14:29:44.745346 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbd72997-3d15-40ee-adfc-b1ffdf7667c9" path="/var/lib/kubelet/pods/dbd72997-3d15-40ee-adfc-b1ffdf7667c9/volumes" Oct 08 14:29:48 crc kubenswrapper[4789]: I1008 14:29:48.730910 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:29:48 crc kubenswrapper[4789]: E1008 14:29:48.732172 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:29:56 crc kubenswrapper[4789]: I1008 14:29:56.038316 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-26f3-account-create-4qzvp"] Oct 08 14:29:56 crc kubenswrapper[4789]: I1008 14:29:56.046506 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-26f3-account-create-4qzvp"] Oct 08 14:29:56 crc kubenswrapper[4789]: I1008 14:29:56.744918 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d43aacf0-f8aa-483b-a0dc-a5b69f9c2984" path="/var/lib/kubelet/pods/d43aacf0-f8aa-483b-a0dc-a5b69f9c2984/volumes" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.170356 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9"] Oct 08 14:30:00 crc kubenswrapper[4789]: E1008 14:30:00.171173 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerName="extract-content" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.171189 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerName="extract-content" Oct 08 14:30:00 crc kubenswrapper[4789]: E1008 14:30:00.171247 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerName="extract-content" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.171261 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerName="extract-content" Oct 08 14:30:00 crc kubenswrapper[4789]: E1008 14:30:00.171278 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerName="registry-server" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.171285 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerName="registry-server" Oct 08 14:30:00 crc kubenswrapper[4789]: E1008 14:30:00.171312 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerName="extract-utilities" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.171318 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerName="extract-utilities" Oct 08 14:30:00 crc kubenswrapper[4789]: E1008 14:30:00.171329 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerName="registry-server" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.171335 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerName="registry-server" Oct 08 14:30:00 crc kubenswrapper[4789]: E1008 14:30:00.171348 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerName="extract-utilities" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.171356 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerName="extract-utilities" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.171645 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e9bb868-6cb7-475d-93c6-2e4306eb7ee2" containerName="registry-server" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.171676 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e76522-68cc-43e7-9811-7ed7441fefc3" containerName="registry-server" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.183585 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9"] Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.183722 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.186005 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.186876 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.347422 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/174b6cc4-86fa-405f-897f-a62690c145fd-config-volume\") pod \"collect-profiles-29332230-gx5c9\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.347478 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h484\" (UniqueName: \"kubernetes.io/projected/174b6cc4-86fa-405f-897f-a62690c145fd-kube-api-access-7h484\") pod \"collect-profiles-29332230-gx5c9\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.347696 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/174b6cc4-86fa-405f-897f-a62690c145fd-secret-volume\") pod \"collect-profiles-29332230-gx5c9\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.449252 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/174b6cc4-86fa-405f-897f-a62690c145fd-config-volume\") pod \"collect-profiles-29332230-gx5c9\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.449323 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h484\" (UniqueName: \"kubernetes.io/projected/174b6cc4-86fa-405f-897f-a62690c145fd-kube-api-access-7h484\") pod \"collect-profiles-29332230-gx5c9\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.449497 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/174b6cc4-86fa-405f-897f-a62690c145fd-secret-volume\") pod \"collect-profiles-29332230-gx5c9\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.450263 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/174b6cc4-86fa-405f-897f-a62690c145fd-config-volume\") pod \"collect-profiles-29332230-gx5c9\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.455538 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/174b6cc4-86fa-405f-897f-a62690c145fd-secret-volume\") pod \"collect-profiles-29332230-gx5c9\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.465930 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h484\" (UniqueName: \"kubernetes.io/projected/174b6cc4-86fa-405f-897f-a62690c145fd-kube-api-access-7h484\") pod \"collect-profiles-29332230-gx5c9\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.510332 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:00 crc kubenswrapper[4789]: I1008 14:30:00.940145 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9"] Oct 08 14:30:01 crc kubenswrapper[4789]: I1008 14:30:01.663885 4789 generic.go:334] "Generic (PLEG): container finished" podID="174b6cc4-86fa-405f-897f-a62690c145fd" containerID="6c6a946f8166584d8b2beb3c190594b70ced2cf42ebef4357e4b1e04c12680bf" exitCode=0 Oct 08 14:30:01 crc kubenswrapper[4789]: I1008 14:30:01.664021 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" event={"ID":"174b6cc4-86fa-405f-897f-a62690c145fd","Type":"ContainerDied","Data":"6c6a946f8166584d8b2beb3c190594b70ced2cf42ebef4357e4b1e04c12680bf"} Oct 08 14:30:01 crc kubenswrapper[4789]: I1008 14:30:01.664227 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" event={"ID":"174b6cc4-86fa-405f-897f-a62690c145fd","Type":"ContainerStarted","Data":"81aec946de14f27e9bd9d8102f0879f96211180082eb953275418869a888436c"} Oct 08 14:30:02 crc kubenswrapper[4789]: I1008 14:30:02.730682 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:30:02 crc kubenswrapper[4789]: E1008 14:30:02.731549 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.016511 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.040748 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-xjfkl"] Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.052815 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-xjfkl"] Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.097578 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/174b6cc4-86fa-405f-897f-a62690c145fd-config-volume\") pod \"174b6cc4-86fa-405f-897f-a62690c145fd\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.097683 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h484\" (UniqueName: \"kubernetes.io/projected/174b6cc4-86fa-405f-897f-a62690c145fd-kube-api-access-7h484\") pod \"174b6cc4-86fa-405f-897f-a62690c145fd\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.097820 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/174b6cc4-86fa-405f-897f-a62690c145fd-secret-volume\") pod \"174b6cc4-86fa-405f-897f-a62690c145fd\" (UID: \"174b6cc4-86fa-405f-897f-a62690c145fd\") " Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.098208 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/174b6cc4-86fa-405f-897f-a62690c145fd-config-volume" (OuterVolumeSpecName: "config-volume") pod "174b6cc4-86fa-405f-897f-a62690c145fd" (UID: "174b6cc4-86fa-405f-897f-a62690c145fd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.103542 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/174b6cc4-86fa-405f-897f-a62690c145fd-kube-api-access-7h484" (OuterVolumeSpecName: "kube-api-access-7h484") pod "174b6cc4-86fa-405f-897f-a62690c145fd" (UID: "174b6cc4-86fa-405f-897f-a62690c145fd"). InnerVolumeSpecName "kube-api-access-7h484". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.104238 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/174b6cc4-86fa-405f-897f-a62690c145fd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "174b6cc4-86fa-405f-897f-a62690c145fd" (UID: "174b6cc4-86fa-405f-897f-a62690c145fd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.200095 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/174b6cc4-86fa-405f-897f-a62690c145fd-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.200127 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/174b6cc4-86fa-405f-897f-a62690c145fd-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.200137 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h484\" (UniqueName: \"kubernetes.io/projected/174b6cc4-86fa-405f-897f-a62690c145fd-kube-api-access-7h484\") on node \"crc\" DevicePath \"\"" Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.683511 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" event={"ID":"174b6cc4-86fa-405f-897f-a62690c145fd","Type":"ContainerDied","Data":"81aec946de14f27e9bd9d8102f0879f96211180082eb953275418869a888436c"} Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.683560 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81aec946de14f27e9bd9d8102f0879f96211180082eb953275418869a888436c" Oct 08 14:30:03 crc kubenswrapper[4789]: I1008 14:30:03.683562 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9" Oct 08 14:30:04 crc kubenswrapper[4789]: I1008 14:30:04.741187 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48f2731e-7573-4376-96bb-e19ac616a706" path="/var/lib/kubelet/pods/48f2731e-7573-4376-96bb-e19ac616a706/volumes" Oct 08 14:30:05 crc kubenswrapper[4789]: I1008 14:30:05.023260 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4fhfk"] Oct 08 14:30:05 crc kubenswrapper[4789]: I1008 14:30:05.032173 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4fhfk"] Oct 08 14:30:06 crc kubenswrapper[4789]: I1008 14:30:06.744446 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e2b82e2-a807-49f8-9c77-f12a8b8bde3c" path="/var/lib/kubelet/pods/8e2b82e2-a807-49f8-9c77-f12a8b8bde3c/volumes" Oct 08 14:30:08 crc kubenswrapper[4789]: I1008 14:30:08.029568 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-9wtgk"] Oct 08 14:30:08 crc kubenswrapper[4789]: I1008 14:30:08.038132 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-9wtgk"] Oct 08 14:30:08 crc kubenswrapper[4789]: I1008 14:30:08.742133 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ded6af-86ab-4942-b141-14db479a5986" path="/var/lib/kubelet/pods/19ded6af-86ab-4942-b141-14db479a5986/volumes" Oct 08 14:30:16 crc kubenswrapper[4789]: I1008 14:30:16.026143 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-7nqh5"] Oct 08 14:30:16 crc kubenswrapper[4789]: I1008 14:30:16.035190 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-7nqh5"] Oct 08 14:30:16 crc kubenswrapper[4789]: I1008 14:30:16.730519 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:30:16 crc kubenswrapper[4789]: E1008 14:30:16.731002 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:30:16 crc kubenswrapper[4789]: I1008 14:30:16.744716 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a038fc03-8ce7-4641-a8f1-1e2a748847a9" path="/var/lib/kubelet/pods/a038fc03-8ce7-4641-a8f1-1e2a748847a9/volumes" Oct 08 14:30:20 crc kubenswrapper[4789]: I1008 14:30:20.041972 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-e02c-account-create-zrxdw"] Oct 08 14:30:20 crc kubenswrapper[4789]: I1008 14:30:20.053254 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-1f09-account-create-4n6fh"] Oct 08 14:30:20 crc kubenswrapper[4789]: I1008 14:30:20.061073 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-1f09-account-create-4n6fh"] Oct 08 14:30:20 crc kubenswrapper[4789]: I1008 14:30:20.068910 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-e02c-account-create-zrxdw"] Oct 08 14:30:20 crc kubenswrapper[4789]: I1008 14:30:20.742575 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a6157f7-df72-4d4a-a3fb-b928c430b4ee" path="/var/lib/kubelet/pods/4a6157f7-df72-4d4a-a3fb-b928c430b4ee/volumes" Oct 08 14:30:20 crc kubenswrapper[4789]: I1008 14:30:20.743664 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd52452e-0042-42ae-9c01-649d38f32965" path="/var/lib/kubelet/pods/fd52452e-0042-42ae-9c01-649d38f32965/volumes" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.154926 4789 scope.go:117] "RemoveContainer" containerID="2362393d2ec81bda1d6fb6b193b12618151d9c80f38bc6f937db8e07b6acdd1b" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.177783 4789 scope.go:117] "RemoveContainer" containerID="7c5cab1656774270d9ae7b1c3f873d33d7a50c423794b564e1e73089bf5a9731" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.221931 4789 scope.go:117] "RemoveContainer" containerID="2e3aa0666eef15cf20b39327a33e2f1c526f956d6c495f0f19590c680f861dc4" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.263173 4789 scope.go:117] "RemoveContainer" containerID="efd39544356733c4b6909daa8d693bbe77fd4236b16e83b4cb22ea18e07587a0" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.313646 4789 scope.go:117] "RemoveContainer" containerID="e2a9d30875924c891b81d7e19dd1b2f459d6c0a4e8b702e049efaf70f4cc4f38" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.378693 4789 scope.go:117] "RemoveContainer" containerID="f87b12863adf8ed4999fe7348da7b1619198540ccdb079fedb1a3afe3af081b3" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.429387 4789 scope.go:117] "RemoveContainer" containerID="4988a5e1718962ab0e50dc3e6bd1fd6823f07a8bd11141afebb5f985202e9af4" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.451514 4789 scope.go:117] "RemoveContainer" containerID="62ded9631a4b6121d80f881faf6895ef2d0c3dd4da082bd8cd584d3828f5d23d" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.473875 4789 scope.go:117] "RemoveContainer" containerID="71929a9f9c5be53fcac8c7aaec81ee8fd28dc08fd3116ec63a85ba6d363edeca" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.492723 4789 scope.go:117] "RemoveContainer" containerID="905c5315d8f00d0d37d0da6d4ca72b638315f50d6e6f45fa0a0b31a0f3c00ed4" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.515389 4789 scope.go:117] "RemoveContainer" containerID="d99c1ff3eee5ae97cd5a41fd82ae1583be6522fe8c2476158488d5691c0d3b16" Oct 08 14:30:21 crc kubenswrapper[4789]: I1008 14:30:21.545571 4789 scope.go:117] "RemoveContainer" containerID="8ad7e6f23e8f8181d1905ac0b159f23d29a075cb04967e012a3e50d3819aba18" Oct 08 14:30:26 crc kubenswrapper[4789]: I1008 14:30:26.031097 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f16b-account-create-vkgg2"] Oct 08 14:30:26 crc kubenswrapper[4789]: I1008 14:30:26.038674 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-266d-account-create-tc8nw"] Oct 08 14:30:26 crc kubenswrapper[4789]: I1008 14:30:26.047141 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f16b-account-create-vkgg2"] Oct 08 14:30:26 crc kubenswrapper[4789]: I1008 14:30:26.055012 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-266d-account-create-tc8nw"] Oct 08 14:30:26 crc kubenswrapper[4789]: I1008 14:30:26.740270 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a7857f4-1cb7-4aed-b42f-401f54afb6c8" path="/var/lib/kubelet/pods/1a7857f4-1cb7-4aed-b42f-401f54afb6c8/volumes" Oct 08 14:30:26 crc kubenswrapper[4789]: I1008 14:30:26.741281 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52156123-e5c0-4e28-b9dc-f972398de562" path="/var/lib/kubelet/pods/52156123-e5c0-4e28-b9dc-f972398de562/volumes" Oct 08 14:30:29 crc kubenswrapper[4789]: I1008 14:30:29.729580 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:30:29 crc kubenswrapper[4789]: E1008 14:30:29.731089 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:30:39 crc kubenswrapper[4789]: I1008 14:30:39.035271 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-srh88"] Oct 08 14:30:39 crc kubenswrapper[4789]: I1008 14:30:39.044177 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-srh88"] Oct 08 14:30:40 crc kubenswrapper[4789]: I1008 14:30:40.746244 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="187c395c-7b8d-45e7-b96c-a93cea38b025" path="/var/lib/kubelet/pods/187c395c-7b8d-45e7-b96c-a93cea38b025/volumes" Oct 08 14:30:43 crc kubenswrapper[4789]: I1008 14:30:43.730417 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:30:43 crc kubenswrapper[4789]: E1008 14:30:43.731518 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:30:49 crc kubenswrapper[4789]: I1008 14:30:49.045536 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-6w7kw"] Oct 08 14:30:49 crc kubenswrapper[4789]: I1008 14:30:49.058283 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-6w7kw"] Oct 08 14:30:50 crc kubenswrapper[4789]: I1008 14:30:50.740301 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c629d8f6-586d-4f5c-a7ea-2e6912a02b03" path="/var/lib/kubelet/pods/c629d8f6-586d-4f5c-a7ea-2e6912a02b03/volumes" Oct 08 14:30:58 crc kubenswrapper[4789]: I1008 14:30:58.730118 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:30:58 crc kubenswrapper[4789]: E1008 14:30:58.731019 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:31:07 crc kubenswrapper[4789]: I1008 14:31:07.291233 4789 generic.go:334] "Generic (PLEG): container finished" podID="5f34622a-a071-45a1-b7fd-75fa8fbb70fe" containerID="f4a6021200876b90bc644aab857483db581ee64d37314bbbd56442cc9cfb8bf4" exitCode=0 Oct 08 14:31:07 crc kubenswrapper[4789]: I1008 14:31:07.291299 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" event={"ID":"5f34622a-a071-45a1-b7fd-75fa8fbb70fe","Type":"ContainerDied","Data":"f4a6021200876b90bc644aab857483db581ee64d37314bbbd56442cc9cfb8bf4"} Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.701866 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.818630 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2shjp\" (UniqueName: \"kubernetes.io/projected/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-kube-api-access-2shjp\") pod \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.818720 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-ssh-key\") pod \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.818918 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-inventory\") pod \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.819052 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-bootstrap-combined-ca-bundle\") pod \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\" (UID: \"5f34622a-a071-45a1-b7fd-75fa8fbb70fe\") " Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.824163 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-kube-api-access-2shjp" (OuterVolumeSpecName: "kube-api-access-2shjp") pod "5f34622a-a071-45a1-b7fd-75fa8fbb70fe" (UID: "5f34622a-a071-45a1-b7fd-75fa8fbb70fe"). InnerVolumeSpecName "kube-api-access-2shjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.824535 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5f34622a-a071-45a1-b7fd-75fa8fbb70fe" (UID: "5f34622a-a071-45a1-b7fd-75fa8fbb70fe"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.846733 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-inventory" (OuterVolumeSpecName: "inventory") pod "5f34622a-a071-45a1-b7fd-75fa8fbb70fe" (UID: "5f34622a-a071-45a1-b7fd-75fa8fbb70fe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.848170 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f34622a-a071-45a1-b7fd-75fa8fbb70fe" (UID: "5f34622a-a071-45a1-b7fd-75fa8fbb70fe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.922313 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2shjp\" (UniqueName: \"kubernetes.io/projected/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-kube-api-access-2shjp\") on node \"crc\" DevicePath \"\"" Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.922348 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.922359 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:31:08 crc kubenswrapper[4789]: I1008 14:31:08.922371 4789 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f34622a-a071-45a1-b7fd-75fa8fbb70fe-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.312079 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" event={"ID":"5f34622a-a071-45a1-b7fd-75fa8fbb70fe","Type":"ContainerDied","Data":"1a56b1f3a0c82b78fae88d73abedef4849e9c59f4608915fac2854d1332d4afa"} Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.312131 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a56b1f3a0c82b78fae88d73abedef4849e9c59f4608915fac2854d1332d4afa" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.312173 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.435160 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd"] Oct 08 14:31:09 crc kubenswrapper[4789]: E1008 14:31:09.435606 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f34622a-a071-45a1-b7fd-75fa8fbb70fe" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.435623 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f34622a-a071-45a1-b7fd-75fa8fbb70fe" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 14:31:09 crc kubenswrapper[4789]: E1008 14:31:09.435636 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="174b6cc4-86fa-405f-897f-a62690c145fd" containerName="collect-profiles" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.435642 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="174b6cc4-86fa-405f-897f-a62690c145fd" containerName="collect-profiles" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.435824 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f34622a-a071-45a1-b7fd-75fa8fbb70fe" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.435841 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="174b6cc4-86fa-405f-897f-a62690c145fd" containerName="collect-profiles" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.436652 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.440155 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.440753 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.440869 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.441883 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.453045 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd"] Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.534512 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.534635 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.534713 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvhw7\" (UniqueName: \"kubernetes.io/projected/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-kube-api-access-dvhw7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.636538 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvhw7\" (UniqueName: \"kubernetes.io/projected/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-kube-api-access-dvhw7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.636634 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.636712 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.640932 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.646264 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.654735 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvhw7\" (UniqueName: \"kubernetes.io/projected/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-kube-api-access-dvhw7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:09 crc kubenswrapper[4789]: I1008 14:31:09.751337 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:31:10 crc kubenswrapper[4789]: I1008 14:31:10.301626 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd"] Oct 08 14:31:10 crc kubenswrapper[4789]: I1008 14:31:10.326197 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" event={"ID":"617364bf-6612-48bf-bfb4-0eeaf9ff2c60","Type":"ContainerStarted","Data":"67951d726ed96a6e9ed1a5c55e3c017db7f1b2264e5823436193452d1c9c23ba"} Oct 08 14:31:11 crc kubenswrapper[4789]: I1008 14:31:11.336261 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" event={"ID":"617364bf-6612-48bf-bfb4-0eeaf9ff2c60","Type":"ContainerStarted","Data":"b3b4e2c131fe9ae0c967daf504eb7d08efaa5b06b0af97101ae3e0e8d4b49c06"} Oct 08 14:31:11 crc kubenswrapper[4789]: I1008 14:31:11.350445 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" podStartSLOduration=1.780744593 podStartE2EDuration="2.350425264s" podCreationTimestamp="2025-10-08 14:31:09 +0000 UTC" firstStartedPulling="2025-10-08 14:31:10.305414811 +0000 UTC m=+1810.212162303" lastFinishedPulling="2025-10-08 14:31:10.875095482 +0000 UTC m=+1810.781842974" observedRunningTime="2025-10-08 14:31:11.350204448 +0000 UTC m=+1811.256951950" watchObservedRunningTime="2025-10-08 14:31:11.350425264 +0000 UTC m=+1811.257172766" Oct 08 14:31:13 crc kubenswrapper[4789]: I1008 14:31:13.730649 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:31:13 crc kubenswrapper[4789]: E1008 14:31:13.731239 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:31:21 crc kubenswrapper[4789]: I1008 14:31:21.826588 4789 scope.go:117] "RemoveContainer" containerID="f3b0624ff4e9bf2a1eabcc055368c6411157ea0b7f80ecb26bf60238681b483d" Oct 08 14:31:21 crc kubenswrapper[4789]: I1008 14:31:21.850357 4789 scope.go:117] "RemoveContainer" containerID="54d9d255673843e618360f4e57b1dfe2bfe4e54a8bf1e14f224ed47b041609cc" Oct 08 14:31:21 crc kubenswrapper[4789]: I1008 14:31:21.888978 4789 scope.go:117] "RemoveContainer" containerID="fd24d64491398970d54cdd13f113442754d73038f8bab41f05c197728fe01471" Oct 08 14:31:21 crc kubenswrapper[4789]: I1008 14:31:21.925452 4789 scope.go:117] "RemoveContainer" containerID="6c56f8c678f13de06c504d4b89c16ed3c33342a592c3dd2df0c28de159d97fc7" Oct 08 14:31:21 crc kubenswrapper[4789]: I1008 14:31:21.962143 4789 scope.go:117] "RemoveContainer" containerID="ca581b859af141bd3b74181028f8be51e086d09b60a46876a0da0c1d4c16ba8b" Oct 08 14:31:22 crc kubenswrapper[4789]: I1008 14:31:22.008325 4789 scope.go:117] "RemoveContainer" containerID="76a8ccf7019532e6e1c615e7262b05ef49e6789012cf931985df04ad37dea96a" Oct 08 14:31:25 crc kubenswrapper[4789]: I1008 14:31:25.729676 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:31:25 crc kubenswrapper[4789]: E1008 14:31:25.730535 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:31:38 crc kubenswrapper[4789]: I1008 14:31:38.730809 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:31:38 crc kubenswrapper[4789]: E1008 14:31:38.732908 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:31:45 crc kubenswrapper[4789]: I1008 14:31:45.059220 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-msj9f"] Oct 08 14:31:45 crc kubenswrapper[4789]: I1008 14:31:45.070069 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-msj9f"] Oct 08 14:31:46 crc kubenswrapper[4789]: I1008 14:31:46.744495 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f21f66b-b493-409d-aee9-fe6b97b468f5" path="/var/lib/kubelet/pods/5f21f66b-b493-409d-aee9-fe6b97b468f5/volumes" Oct 08 14:31:50 crc kubenswrapper[4789]: I1008 14:31:50.735707 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:31:50 crc kubenswrapper[4789]: E1008 14:31:50.736248 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:32:04 crc kubenswrapper[4789]: I1008 14:32:04.730597 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:32:05 crc kubenswrapper[4789]: I1008 14:32:05.805972 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"92b1a760457dbd1ac1a6a96f521ea2565954a40aa5a1f3f3fff338d563639227"} Oct 08 14:32:22 crc kubenswrapper[4789]: I1008 14:32:22.214881 4789 scope.go:117] "RemoveContainer" containerID="ef28cc9775c296870b3049d94d755a39df9b707fd45dc30f2dff0db73f8de5d4" Oct 08 14:32:28 crc kubenswrapper[4789]: I1008 14:32:28.047921 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-flz9d"] Oct 08 14:32:28 crc kubenswrapper[4789]: I1008 14:32:28.057560 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-flz9d"] Oct 08 14:32:28 crc kubenswrapper[4789]: I1008 14:32:28.753175 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9" path="/var/lib/kubelet/pods/ac7cd0ec-3f8d-4afa-84de-cfbb47a7fab9/volumes" Oct 08 14:32:52 crc kubenswrapper[4789]: I1008 14:32:52.039478 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-hks4r"] Oct 08 14:32:52 crc kubenswrapper[4789]: I1008 14:32:52.051841 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-hks4r"] Oct 08 14:32:52 crc kubenswrapper[4789]: I1008 14:32:52.741298 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dccf34c5-6e31-4d04-9478-d4e4edcc6ab0" path="/var/lib/kubelet/pods/dccf34c5-6e31-4d04-9478-d4e4edcc6ab0/volumes" Oct 08 14:32:55 crc kubenswrapper[4789]: I1008 14:32:55.030620 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-tblv4"] Oct 08 14:32:55 crc kubenswrapper[4789]: I1008 14:32:55.038485 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-tblv4"] Oct 08 14:32:56 crc kubenswrapper[4789]: I1008 14:32:56.749306 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef2b1def-1464-4d5e-8d4e-44a4e505fa40" path="/var/lib/kubelet/pods/ef2b1def-1464-4d5e-8d4e-44a4e505fa40/volumes" Oct 08 14:33:09 crc kubenswrapper[4789]: I1008 14:33:09.026825 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-xll5s"] Oct 08 14:33:09 crc kubenswrapper[4789]: I1008 14:33:09.035440 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-xll5s"] Oct 08 14:33:10 crc kubenswrapper[4789]: I1008 14:33:10.027489 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-69jjr"] Oct 08 14:33:10 crc kubenswrapper[4789]: I1008 14:33:10.038474 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-2wfw9"] Oct 08 14:33:10 crc kubenswrapper[4789]: I1008 14:33:10.048809 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-69jjr"] Oct 08 14:33:10 crc kubenswrapper[4789]: I1008 14:33:10.058651 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-2wfw9"] Oct 08 14:33:10 crc kubenswrapper[4789]: I1008 14:33:10.399398 4789 generic.go:334] "Generic (PLEG): container finished" podID="617364bf-6612-48bf-bfb4-0eeaf9ff2c60" containerID="b3b4e2c131fe9ae0c967daf504eb7d08efaa5b06b0af97101ae3e0e8d4b49c06" exitCode=0 Oct 08 14:33:10 crc kubenswrapper[4789]: I1008 14:33:10.399448 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" event={"ID":"617364bf-6612-48bf-bfb4-0eeaf9ff2c60","Type":"ContainerDied","Data":"b3b4e2c131fe9ae0c967daf504eb7d08efaa5b06b0af97101ae3e0e8d4b49c06"} Oct 08 14:33:10 crc kubenswrapper[4789]: I1008 14:33:10.741833 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd" path="/var/lib/kubelet/pods/01e9678e-d1ed-4a11-9c99-ddfc8cb5e4dd/volumes" Oct 08 14:33:10 crc kubenswrapper[4789]: I1008 14:33:10.744240 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05a2fc22-aea2-42aa-810a-98bbf5ba7c98" path="/var/lib/kubelet/pods/05a2fc22-aea2-42aa-810a-98bbf5ba7c98/volumes" Oct 08 14:33:10 crc kubenswrapper[4789]: I1008 14:33:10.746174 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="150d7f22-fe1e-47cf-bf5e-7f97473446b8" path="/var/lib/kubelet/pods/150d7f22-fe1e-47cf-bf5e-7f97473446b8/volumes" Oct 08 14:33:11 crc kubenswrapper[4789]: I1008 14:33:11.863656 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.008456 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-ssh-key\") pod \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.008558 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-inventory\") pod \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.008593 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvhw7\" (UniqueName: \"kubernetes.io/projected/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-kube-api-access-dvhw7\") pod \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\" (UID: \"617364bf-6612-48bf-bfb4-0eeaf9ff2c60\") " Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.020157 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-kube-api-access-dvhw7" (OuterVolumeSpecName: "kube-api-access-dvhw7") pod "617364bf-6612-48bf-bfb4-0eeaf9ff2c60" (UID: "617364bf-6612-48bf-bfb4-0eeaf9ff2c60"). InnerVolumeSpecName "kube-api-access-dvhw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.040086 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-inventory" (OuterVolumeSpecName: "inventory") pod "617364bf-6612-48bf-bfb4-0eeaf9ff2c60" (UID: "617364bf-6612-48bf-bfb4-0eeaf9ff2c60"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.040766 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "617364bf-6612-48bf-bfb4-0eeaf9ff2c60" (UID: "617364bf-6612-48bf-bfb4-0eeaf9ff2c60"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.110574 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.110611 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.110625 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvhw7\" (UniqueName: \"kubernetes.io/projected/617364bf-6612-48bf-bfb4-0eeaf9ff2c60-kube-api-access-dvhw7\") on node \"crc\" DevicePath \"\"" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.421602 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" event={"ID":"617364bf-6612-48bf-bfb4-0eeaf9ff2c60","Type":"ContainerDied","Data":"67951d726ed96a6e9ed1a5c55e3c017db7f1b2264e5823436193452d1c9c23ba"} Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.421664 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67951d726ed96a6e9ed1a5c55e3c017db7f1b2264e5823436193452d1c9c23ba" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.421755 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.513473 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9"] Oct 08 14:33:12 crc kubenswrapper[4789]: E1008 14:33:12.514239 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="617364bf-6612-48bf-bfb4-0eeaf9ff2c60" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.514252 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="617364bf-6612-48bf-bfb4-0eeaf9ff2c60" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.514465 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="617364bf-6612-48bf-bfb4-0eeaf9ff2c60" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.515269 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.519359 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.519436 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.522158 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.522346 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.526011 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9"] Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.620016 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.620170 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.620520 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlt82\" (UniqueName: \"kubernetes.io/projected/48dac93b-b8ea-4016-9a48-36fd0c594bf4-kube-api-access-zlt82\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.722205 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlt82\" (UniqueName: \"kubernetes.io/projected/48dac93b-b8ea-4016-9a48-36fd0c594bf4-kube-api-access-zlt82\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.722367 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.722446 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.726669 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.730558 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.739411 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlt82\" (UniqueName: \"kubernetes.io/projected/48dac93b-b8ea-4016-9a48-36fd0c594bf4-kube-api-access-zlt82\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:12 crc kubenswrapper[4789]: I1008 14:33:12.841228 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:33:13 crc kubenswrapper[4789]: I1008 14:33:13.042287 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-5c9mc"] Oct 08 14:33:13 crc kubenswrapper[4789]: I1008 14:33:13.054394 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-5c9mc"] Oct 08 14:33:13 crc kubenswrapper[4789]: I1008 14:33:13.333095 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9"] Oct 08 14:33:13 crc kubenswrapper[4789]: I1008 14:33:13.336187 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 14:33:13 crc kubenswrapper[4789]: I1008 14:33:13.433378 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" event={"ID":"48dac93b-b8ea-4016-9a48-36fd0c594bf4","Type":"ContainerStarted","Data":"268b1c8731891165f49b0dee7c32c099c7d8f7bfa278d1751c1d5bc87215931d"} Oct 08 14:33:14 crc kubenswrapper[4789]: I1008 14:33:14.028866 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-p54dp"] Oct 08 14:33:14 crc kubenswrapper[4789]: I1008 14:33:14.036970 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-p54dp"] Oct 08 14:33:14 crc kubenswrapper[4789]: I1008 14:33:14.758428 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b096b2c-0d8d-44bc-9275-e712b6e622b8" path="/var/lib/kubelet/pods/0b096b2c-0d8d-44bc-9275-e712b6e622b8/volumes" Oct 08 14:33:14 crc kubenswrapper[4789]: I1008 14:33:14.761302 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d62cf1d-f82b-4eed-8500-2614367a4e3d" path="/var/lib/kubelet/pods/5d62cf1d-f82b-4eed-8500-2614367a4e3d/volumes" Oct 08 14:33:16 crc kubenswrapper[4789]: I1008 14:33:16.461866 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" event={"ID":"48dac93b-b8ea-4016-9a48-36fd0c594bf4","Type":"ContainerStarted","Data":"5a55853cf0f1aaad1ee19ff04044b6b68ccb9856d0bb386ffd966960a8a69dee"} Oct 08 14:33:16 crc kubenswrapper[4789]: I1008 14:33:16.485593 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" podStartSLOduration=2.098659627 podStartE2EDuration="4.48557568s" podCreationTimestamp="2025-10-08 14:33:12 +0000 UTC" firstStartedPulling="2025-10-08 14:33:13.335887866 +0000 UTC m=+1933.242635358" lastFinishedPulling="2025-10-08 14:33:15.722803909 +0000 UTC m=+1935.629551411" observedRunningTime="2025-10-08 14:33:16.480788709 +0000 UTC m=+1936.387536201" watchObservedRunningTime="2025-10-08 14:33:16.48557568 +0000 UTC m=+1936.392323172" Oct 08 14:33:20 crc kubenswrapper[4789]: I1008 14:33:20.060069 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-e461-account-create-sgqdc"] Oct 08 14:33:20 crc kubenswrapper[4789]: I1008 14:33:20.073623 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-b687-account-create-4s2zx"] Oct 08 14:33:20 crc kubenswrapper[4789]: I1008 14:33:20.080721 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-e461-account-create-sgqdc"] Oct 08 14:33:20 crc kubenswrapper[4789]: I1008 14:33:20.087647 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-b687-account-create-4s2zx"] Oct 08 14:33:20 crc kubenswrapper[4789]: I1008 14:33:20.740648 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36757dfa-8de4-475b-bf51-7320d55c16c7" path="/var/lib/kubelet/pods/36757dfa-8de4-475b-bf51-7320d55c16c7/volumes" Oct 08 14:33:20 crc kubenswrapper[4789]: I1008 14:33:20.742174 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b985f8e-1745-4c46-9f99-d1e4978413e0" path="/var/lib/kubelet/pods/9b985f8e-1745-4c46-9f99-d1e4978413e0/volumes" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.295246 4789 scope.go:117] "RemoveContainer" containerID="96a413984afa6267b4352c3740ff023d52b3ca71b3f4a31c97e9b86f7c99941c" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.357730 4789 scope.go:117] "RemoveContainer" containerID="c0252b016c5a8f9976248ef4ed38cc13853d5b34265149fb7797f1bf250e2874" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.390349 4789 scope.go:117] "RemoveContainer" containerID="c1bf52812a5d2149fb993cd3676c09b67d4e4fdde81d2f4ee339b9debb591dc0" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.466110 4789 scope.go:117] "RemoveContainer" containerID="edd9d78df92c7ffe45486e262e4ca4142bcaa9ee65566bc9f75c179f44c9d174" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.500188 4789 scope.go:117] "RemoveContainer" containerID="25383f9b103897430455bdc8eb3b2c62de60a49ab00c45d36fae689fc954e7c7" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.565867 4789 scope.go:117] "RemoveContainer" containerID="c6a9a9f5178d5155c28bdee0e913b49dcfa8d0018b023832fff25943780f9225" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.608359 4789 scope.go:117] "RemoveContainer" containerID="88d3dc78c1125ad9e5c403a1c650c6d4011e790180ddbfb07cd5d56050d2a189" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.645543 4789 scope.go:117] "RemoveContainer" containerID="4855c397e082720541e29543832b80834d6c77a86884b6f30d2c64d033f7c2de" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.668335 4789 scope.go:117] "RemoveContainer" containerID="0b0da7215214f8a92093bc4106a2506ee926eac0474f267783e915ef17c69b45" Oct 08 14:33:22 crc kubenswrapper[4789]: I1008 14:33:22.711268 4789 scope.go:117] "RemoveContainer" containerID="933f8f193cc92a17c8ff9a268982d0bc59d0534b608fb4c11e4b1f07def68aa2" Oct 08 14:33:30 crc kubenswrapper[4789]: I1008 14:33:30.023612 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-a3a4-account-create-7wxzg"] Oct 08 14:33:30 crc kubenswrapper[4789]: I1008 14:33:30.030447 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-a3a4-account-create-7wxzg"] Oct 08 14:33:30 crc kubenswrapper[4789]: I1008 14:33:30.743200 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbc3a435-e2f9-4353-910c-b62c505278d7" path="/var/lib/kubelet/pods/bbc3a435-e2f9-4353-910c-b62c505278d7/volumes" Oct 08 14:34:22 crc kubenswrapper[4789]: I1008 14:34:22.941253 4789 scope.go:117] "RemoveContainer" containerID="4af4d007406050d269006b2dfb41b07b88dab2589dd6f8ac0cb2383e58f5d12e" Oct 08 14:34:26 crc kubenswrapper[4789]: I1008 14:34:26.433121 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:34:26 crc kubenswrapper[4789]: I1008 14:34:26.433594 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:34:27 crc kubenswrapper[4789]: I1008 14:34:27.038498 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-75hcm"] Oct 08 14:34:27 crc kubenswrapper[4789]: I1008 14:34:27.048814 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-75hcm"] Oct 08 14:34:28 crc kubenswrapper[4789]: I1008 14:34:28.741503 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4779c61-4a62-4853-b877-fbf588daa8e8" path="/var/lib/kubelet/pods/c4779c61-4a62-4853-b877-fbf588daa8e8/volumes" Oct 08 14:34:33 crc kubenswrapper[4789]: I1008 14:34:33.164621 4789 generic.go:334] "Generic (PLEG): container finished" podID="48dac93b-b8ea-4016-9a48-36fd0c594bf4" containerID="5a55853cf0f1aaad1ee19ff04044b6b68ccb9856d0bb386ffd966960a8a69dee" exitCode=0 Oct 08 14:34:33 crc kubenswrapper[4789]: I1008 14:34:33.164683 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" event={"ID":"48dac93b-b8ea-4016-9a48-36fd0c594bf4","Type":"ContainerDied","Data":"5a55853cf0f1aaad1ee19ff04044b6b68ccb9856d0bb386ffd966960a8a69dee"} Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.583076 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.738157 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-ssh-key\") pod \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.738250 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-inventory\") pod \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.738314 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlt82\" (UniqueName: \"kubernetes.io/projected/48dac93b-b8ea-4016-9a48-36fd0c594bf4-kube-api-access-zlt82\") pod \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\" (UID: \"48dac93b-b8ea-4016-9a48-36fd0c594bf4\") " Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.749275 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48dac93b-b8ea-4016-9a48-36fd0c594bf4-kube-api-access-zlt82" (OuterVolumeSpecName: "kube-api-access-zlt82") pod "48dac93b-b8ea-4016-9a48-36fd0c594bf4" (UID: "48dac93b-b8ea-4016-9a48-36fd0c594bf4"). InnerVolumeSpecName "kube-api-access-zlt82". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.770108 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48dac93b-b8ea-4016-9a48-36fd0c594bf4" (UID: "48dac93b-b8ea-4016-9a48-36fd0c594bf4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.771268 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-inventory" (OuterVolumeSpecName: "inventory") pod "48dac93b-b8ea-4016-9a48-36fd0c594bf4" (UID: "48dac93b-b8ea-4016-9a48-36fd0c594bf4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.840174 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.840210 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48dac93b-b8ea-4016-9a48-36fd0c594bf4-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:34:34 crc kubenswrapper[4789]: I1008 14:34:34.840224 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlt82\" (UniqueName: \"kubernetes.io/projected/48dac93b-b8ea-4016-9a48-36fd0c594bf4-kube-api-access-zlt82\") on node \"crc\" DevicePath \"\"" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.183763 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" event={"ID":"48dac93b-b8ea-4016-9a48-36fd0c594bf4","Type":"ContainerDied","Data":"268b1c8731891165f49b0dee7c32c099c7d8f7bfa278d1751c1d5bc87215931d"} Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.183813 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="268b1c8731891165f49b0dee7c32c099c7d8f7bfa278d1751c1d5bc87215931d" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.183780 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.326942 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj"] Oct 08 14:34:35 crc kubenswrapper[4789]: E1008 14:34:35.327515 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48dac93b-b8ea-4016-9a48-36fd0c594bf4" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.327538 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="48dac93b-b8ea-4016-9a48-36fd0c594bf4" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.327798 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="48dac93b-b8ea-4016-9a48-36fd0c594bf4" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.328610 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.331490 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.331872 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.331915 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.332086 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.338460 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj"] Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.453654 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.453748 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.453810 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq9jg\" (UniqueName: \"kubernetes.io/projected/04be0498-514b-45df-9ff0-7eae07586037-kube-api-access-xq9jg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.555353 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.555444 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq9jg\" (UniqueName: \"kubernetes.io/projected/04be0498-514b-45df-9ff0-7eae07586037-kube-api-access-xq9jg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.555550 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.558680 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.562775 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.571086 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq9jg\" (UniqueName: \"kubernetes.io/projected/04be0498-514b-45df-9ff0-7eae07586037-kube-api-access-xq9jg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:35 crc kubenswrapper[4789]: I1008 14:34:35.650655 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:36 crc kubenswrapper[4789]: I1008 14:34:36.134773 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj"] Oct 08 14:34:36 crc kubenswrapper[4789]: I1008 14:34:36.200750 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" event={"ID":"04be0498-514b-45df-9ff0-7eae07586037","Type":"ContainerStarted","Data":"dae1c5fea82e2828eb4b66bb700b2e2b9cfa652c3508b6fbcaa0d4c6bb6b75f6"} Oct 08 14:34:37 crc kubenswrapper[4789]: I1008 14:34:37.210069 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" event={"ID":"04be0498-514b-45df-9ff0-7eae07586037","Type":"ContainerStarted","Data":"8da2b9da42f140019f6d8751d70c681f1c5be19fcd137a3d943da6e80e2b0bd8"} Oct 08 14:34:37 crc kubenswrapper[4789]: I1008 14:34:37.226930 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" podStartSLOduration=1.597167942 podStartE2EDuration="2.226893045s" podCreationTimestamp="2025-10-08 14:34:35 +0000 UTC" firstStartedPulling="2025-10-08 14:34:36.144821126 +0000 UTC m=+2016.051568608" lastFinishedPulling="2025-10-08 14:34:36.774546219 +0000 UTC m=+2016.681293711" observedRunningTime="2025-10-08 14:34:37.222637598 +0000 UTC m=+2017.129385090" watchObservedRunningTime="2025-10-08 14:34:37.226893045 +0000 UTC m=+2017.133640537" Oct 08 14:34:42 crc kubenswrapper[4789]: I1008 14:34:42.259261 4789 generic.go:334] "Generic (PLEG): container finished" podID="04be0498-514b-45df-9ff0-7eae07586037" containerID="8da2b9da42f140019f6d8751d70c681f1c5be19fcd137a3d943da6e80e2b0bd8" exitCode=0 Oct 08 14:34:42 crc kubenswrapper[4789]: I1008 14:34:42.259306 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" event={"ID":"04be0498-514b-45df-9ff0-7eae07586037","Type":"ContainerDied","Data":"8da2b9da42f140019f6d8751d70c681f1c5be19fcd137a3d943da6e80e2b0bd8"} Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.756885 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.848288 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-ssh-key\") pod \"04be0498-514b-45df-9ff0-7eae07586037\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.848381 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-inventory\") pod \"04be0498-514b-45df-9ff0-7eae07586037\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.848424 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq9jg\" (UniqueName: \"kubernetes.io/projected/04be0498-514b-45df-9ff0-7eae07586037-kube-api-access-xq9jg\") pod \"04be0498-514b-45df-9ff0-7eae07586037\" (UID: \"04be0498-514b-45df-9ff0-7eae07586037\") " Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.854383 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04be0498-514b-45df-9ff0-7eae07586037-kube-api-access-xq9jg" (OuterVolumeSpecName: "kube-api-access-xq9jg") pod "04be0498-514b-45df-9ff0-7eae07586037" (UID: "04be0498-514b-45df-9ff0-7eae07586037"). InnerVolumeSpecName "kube-api-access-xq9jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.885648 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-inventory" (OuterVolumeSpecName: "inventory") pod "04be0498-514b-45df-9ff0-7eae07586037" (UID: "04be0498-514b-45df-9ff0-7eae07586037"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.893031 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "04be0498-514b-45df-9ff0-7eae07586037" (UID: "04be0498-514b-45df-9ff0-7eae07586037"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.950906 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.950945 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04be0498-514b-45df-9ff0-7eae07586037-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:34:43 crc kubenswrapper[4789]: I1008 14:34:43.950957 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq9jg\" (UniqueName: \"kubernetes.io/projected/04be0498-514b-45df-9ff0-7eae07586037-kube-api-access-xq9jg\") on node \"crc\" DevicePath \"\"" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.278486 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" event={"ID":"04be0498-514b-45df-9ff0-7eae07586037","Type":"ContainerDied","Data":"dae1c5fea82e2828eb4b66bb700b2e2b9cfa652c3508b6fbcaa0d4c6bb6b75f6"} Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.278522 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dae1c5fea82e2828eb4b66bb700b2e2b9cfa652c3508b6fbcaa0d4c6bb6b75f6" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.278553 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.355184 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg"] Oct 08 14:34:44 crc kubenswrapper[4789]: E1008 14:34:44.355642 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04be0498-514b-45df-9ff0-7eae07586037" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.355656 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="04be0498-514b-45df-9ff0-7eae07586037" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.355866 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="04be0498-514b-45df-9ff0-7eae07586037" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.356634 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.359242 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.359542 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.359695 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.359889 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.367821 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg"] Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.468954 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xpffg\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.469413 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbcpn\" (UniqueName: \"kubernetes.io/projected/5a8f80d9-3b33-479c-9d6a-77867520218d-kube-api-access-hbcpn\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xpffg\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.469558 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xpffg\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.571881 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbcpn\" (UniqueName: \"kubernetes.io/projected/5a8f80d9-3b33-479c-9d6a-77867520218d-kube-api-access-hbcpn\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xpffg\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.571974 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xpffg\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.572076 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xpffg\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.578735 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xpffg\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.578740 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xpffg\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.589685 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbcpn\" (UniqueName: \"kubernetes.io/projected/5a8f80d9-3b33-479c-9d6a-77867520218d-kube-api-access-hbcpn\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xpffg\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:44 crc kubenswrapper[4789]: I1008 14:34:44.678888 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:34:45 crc kubenswrapper[4789]: I1008 14:34:45.032539 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-v56bp"] Oct 08 14:34:45 crc kubenswrapper[4789]: I1008 14:34:45.039679 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-v56bp"] Oct 08 14:34:45 crc kubenswrapper[4789]: I1008 14:34:45.336730 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg"] Oct 08 14:34:46 crc kubenswrapper[4789]: I1008 14:34:46.298029 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" event={"ID":"5a8f80d9-3b33-479c-9d6a-77867520218d","Type":"ContainerStarted","Data":"83da06805f658a5350d690de0396b6f0156385ca9fbc3ba49b23f111bb725aa5"} Oct 08 14:34:46 crc kubenswrapper[4789]: I1008 14:34:46.298292 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" event={"ID":"5a8f80d9-3b33-479c-9d6a-77867520218d","Type":"ContainerStarted","Data":"e9e1117f101ab3902e396fde109d2a78fb234f8393e944acbebcfd92c481b2ce"} Oct 08 14:34:46 crc kubenswrapper[4789]: I1008 14:34:46.326304 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" podStartSLOduration=1.792791826 podStartE2EDuration="2.326283294s" podCreationTimestamp="2025-10-08 14:34:44 +0000 UTC" firstStartedPulling="2025-10-08 14:34:45.332549164 +0000 UTC m=+2025.239296656" lastFinishedPulling="2025-10-08 14:34:45.866040612 +0000 UTC m=+2025.772788124" observedRunningTime="2025-10-08 14:34:46.31554971 +0000 UTC m=+2026.222297282" watchObservedRunningTime="2025-10-08 14:34:46.326283294 +0000 UTC m=+2026.233030796" Oct 08 14:34:46 crc kubenswrapper[4789]: I1008 14:34:46.739733 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="829d0843-9f3f-4bd4-b5e3-4562c7405f5f" path="/var/lib/kubelet/pods/829d0843-9f3f-4bd4-b5e3-4562c7405f5f/volumes" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.490225 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c9vpr"] Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.502079 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.504663 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9vpr"] Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.683455 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r2n9\" (UniqueName: \"kubernetes.io/projected/de3e4c8e-3195-4ad7-aebe-871b753c35c7-kube-api-access-8r2n9\") pod \"redhat-marketplace-c9vpr\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.683574 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-catalog-content\") pod \"redhat-marketplace-c9vpr\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.683641 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-utilities\") pod \"redhat-marketplace-c9vpr\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.785609 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-utilities\") pod \"redhat-marketplace-c9vpr\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.785712 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r2n9\" (UniqueName: \"kubernetes.io/projected/de3e4c8e-3195-4ad7-aebe-871b753c35c7-kube-api-access-8r2n9\") pod \"redhat-marketplace-c9vpr\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.785792 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-catalog-content\") pod \"redhat-marketplace-c9vpr\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.786183 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-utilities\") pod \"redhat-marketplace-c9vpr\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.786238 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-catalog-content\") pod \"redhat-marketplace-c9vpr\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.810388 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r2n9\" (UniqueName: \"kubernetes.io/projected/de3e4c8e-3195-4ad7-aebe-871b753c35c7-kube-api-access-8r2n9\") pod \"redhat-marketplace-c9vpr\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:50 crc kubenswrapper[4789]: I1008 14:34:50.822338 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:34:51 crc kubenswrapper[4789]: I1008 14:34:51.281839 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9vpr"] Oct 08 14:34:51 crc kubenswrapper[4789]: W1008 14:34:51.287332 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde3e4c8e_3195_4ad7_aebe_871b753c35c7.slice/crio-347b737edd785ed9f6a71e85d9af158b9ad0af8eac55f42b1a91beade2e4d79f WatchSource:0}: Error finding container 347b737edd785ed9f6a71e85d9af158b9ad0af8eac55f42b1a91beade2e4d79f: Status 404 returned error can't find the container with id 347b737edd785ed9f6a71e85d9af158b9ad0af8eac55f42b1a91beade2e4d79f Oct 08 14:34:51 crc kubenswrapper[4789]: I1008 14:34:51.345366 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9vpr" event={"ID":"de3e4c8e-3195-4ad7-aebe-871b753c35c7","Type":"ContainerStarted","Data":"347b737edd785ed9f6a71e85d9af158b9ad0af8eac55f42b1a91beade2e4d79f"} Oct 08 14:34:52 crc kubenswrapper[4789]: I1008 14:34:52.037639 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dr8h4"] Oct 08 14:34:52 crc kubenswrapper[4789]: I1008 14:34:52.048045 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dr8h4"] Oct 08 14:34:52 crc kubenswrapper[4789]: I1008 14:34:52.355806 4789 generic.go:334] "Generic (PLEG): container finished" podID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerID="581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3" exitCode=0 Oct 08 14:34:52 crc kubenswrapper[4789]: I1008 14:34:52.355856 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9vpr" event={"ID":"de3e4c8e-3195-4ad7-aebe-871b753c35c7","Type":"ContainerDied","Data":"581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3"} Oct 08 14:34:52 crc kubenswrapper[4789]: I1008 14:34:52.746758 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aa68379-349e-48f5-b6c5-d2b694639e9f" path="/var/lib/kubelet/pods/4aa68379-349e-48f5-b6c5-d2b694639e9f/volumes" Oct 08 14:34:53 crc kubenswrapper[4789]: I1008 14:34:53.367980 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9vpr" event={"ID":"de3e4c8e-3195-4ad7-aebe-871b753c35c7","Type":"ContainerStarted","Data":"3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8"} Oct 08 14:34:54 crc kubenswrapper[4789]: I1008 14:34:54.378073 4789 generic.go:334] "Generic (PLEG): container finished" podID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerID="3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8" exitCode=0 Oct 08 14:34:54 crc kubenswrapper[4789]: I1008 14:34:54.378117 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9vpr" event={"ID":"de3e4c8e-3195-4ad7-aebe-871b753c35c7","Type":"ContainerDied","Data":"3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8"} Oct 08 14:34:55 crc kubenswrapper[4789]: I1008 14:34:55.388606 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9vpr" event={"ID":"de3e4c8e-3195-4ad7-aebe-871b753c35c7","Type":"ContainerStarted","Data":"4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d"} Oct 08 14:34:55 crc kubenswrapper[4789]: I1008 14:34:55.408245 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c9vpr" podStartSLOduration=2.844443259 podStartE2EDuration="5.408227375s" podCreationTimestamp="2025-10-08 14:34:50 +0000 UTC" firstStartedPulling="2025-10-08 14:34:52.357420819 +0000 UTC m=+2032.264168311" lastFinishedPulling="2025-10-08 14:34:54.921204925 +0000 UTC m=+2034.827952427" observedRunningTime="2025-10-08 14:34:55.402490997 +0000 UTC m=+2035.309238479" watchObservedRunningTime="2025-10-08 14:34:55.408227375 +0000 UTC m=+2035.314974867" Oct 08 14:34:56 crc kubenswrapper[4789]: I1008 14:34:56.433241 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:34:56 crc kubenswrapper[4789]: I1008 14:34:56.433299 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:35:00 crc kubenswrapper[4789]: I1008 14:35:00.823308 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:35:00 crc kubenswrapper[4789]: I1008 14:35:00.828021 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:35:00 crc kubenswrapper[4789]: I1008 14:35:00.870888 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:35:01 crc kubenswrapper[4789]: I1008 14:35:01.537684 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:35:01 crc kubenswrapper[4789]: I1008 14:35:01.581316 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9vpr"] Oct 08 14:35:03 crc kubenswrapper[4789]: I1008 14:35:03.471732 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c9vpr" podUID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerName="registry-server" containerID="cri-o://4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d" gracePeriod=2 Oct 08 14:35:03 crc kubenswrapper[4789]: E1008 14:35:03.907979 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde3e4c8e_3195_4ad7_aebe_871b753c35c7.slice/crio-conmon-4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d.scope\": RecentStats: unable to find data in memory cache]" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.478134 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.492749 4789 generic.go:334] "Generic (PLEG): container finished" podID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerID="4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d" exitCode=0 Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.492815 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9vpr" event={"ID":"de3e4c8e-3195-4ad7-aebe-871b753c35c7","Type":"ContainerDied","Data":"4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d"} Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.492872 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9vpr" event={"ID":"de3e4c8e-3195-4ad7-aebe-871b753c35c7","Type":"ContainerDied","Data":"347b737edd785ed9f6a71e85d9af158b9ad0af8eac55f42b1a91beade2e4d79f"} Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.492899 4789 scope.go:117] "RemoveContainer" containerID="4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.493232 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9vpr" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.524374 4789 scope.go:117] "RemoveContainer" containerID="3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.542326 4789 scope.go:117] "RemoveContainer" containerID="581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.579846 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r2n9\" (UniqueName: \"kubernetes.io/projected/de3e4c8e-3195-4ad7-aebe-871b753c35c7-kube-api-access-8r2n9\") pod \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.580338 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-utilities\") pod \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.580491 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-catalog-content\") pod \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\" (UID: \"de3e4c8e-3195-4ad7-aebe-871b753c35c7\") " Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.581293 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-utilities" (OuterVolumeSpecName: "utilities") pod "de3e4c8e-3195-4ad7-aebe-871b753c35c7" (UID: "de3e4c8e-3195-4ad7-aebe-871b753c35c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.586298 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de3e4c8e-3195-4ad7-aebe-871b753c35c7-kube-api-access-8r2n9" (OuterVolumeSpecName: "kube-api-access-8r2n9") pod "de3e4c8e-3195-4ad7-aebe-871b753c35c7" (UID: "de3e4c8e-3195-4ad7-aebe-871b753c35c7"). InnerVolumeSpecName "kube-api-access-8r2n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.591194 4789 scope.go:117] "RemoveContainer" containerID="4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d" Oct 08 14:35:04 crc kubenswrapper[4789]: E1008 14:35:04.591660 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d\": container with ID starting with 4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d not found: ID does not exist" containerID="4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.591689 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d"} err="failed to get container status \"4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d\": rpc error: code = NotFound desc = could not find container \"4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d\": container with ID starting with 4f1445ec89d4a81ce60be61ba360bc849407a1023d8152bb924afccfdaa0211d not found: ID does not exist" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.591711 4789 scope.go:117] "RemoveContainer" containerID="3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8" Oct 08 14:35:04 crc kubenswrapper[4789]: E1008 14:35:04.591963 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8\": container with ID starting with 3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8 not found: ID does not exist" containerID="3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.592136 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8"} err="failed to get container status \"3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8\": rpc error: code = NotFound desc = could not find container \"3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8\": container with ID starting with 3f18d45c968e89e4f5bbe626562253e9ccb619c9400a073d4903da155ae450e8 not found: ID does not exist" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.592271 4789 scope.go:117] "RemoveContainer" containerID="581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3" Oct 08 14:35:04 crc kubenswrapper[4789]: E1008 14:35:04.592638 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3\": container with ID starting with 581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3 not found: ID does not exist" containerID="581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.592675 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3"} err="failed to get container status \"581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3\": rpc error: code = NotFound desc = could not find container \"581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3\": container with ID starting with 581ac438abd4b5e573c1d7768ecbd2ab6491aedf7ba7b596ed98410524e66ee3 not found: ID does not exist" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.593371 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de3e4c8e-3195-4ad7-aebe-871b753c35c7" (UID: "de3e4c8e-3195-4ad7-aebe-871b753c35c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.683104 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.683130 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de3e4c8e-3195-4ad7-aebe-871b753c35c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.683141 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r2n9\" (UniqueName: \"kubernetes.io/projected/de3e4c8e-3195-4ad7-aebe-871b753c35c7-kube-api-access-8r2n9\") on node \"crc\" DevicePath \"\"" Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.826269 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9vpr"] Oct 08 14:35:04 crc kubenswrapper[4789]: I1008 14:35:04.834231 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9vpr"] Oct 08 14:35:06 crc kubenswrapper[4789]: I1008 14:35:06.743943 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" path="/var/lib/kubelet/pods/de3e4c8e-3195-4ad7-aebe-871b753c35c7/volumes" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.223385 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wnmzs"] Oct 08 14:35:09 crc kubenswrapper[4789]: E1008 14:35:09.224189 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerName="registry-server" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.224226 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerName="registry-server" Oct 08 14:35:09 crc kubenswrapper[4789]: E1008 14:35:09.224255 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerName="extract-utilities" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.224267 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerName="extract-utilities" Oct 08 14:35:09 crc kubenswrapper[4789]: E1008 14:35:09.224293 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerName="extract-content" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.224304 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerName="extract-content" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.224583 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3e4c8e-3195-4ad7-aebe-871b753c35c7" containerName="registry-server" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.229663 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.252364 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wnmzs"] Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.381798 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-catalog-content\") pod \"community-operators-wnmzs\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.381907 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-utilities\") pod \"community-operators-wnmzs\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.382328 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64bgl\" (UniqueName: \"kubernetes.io/projected/f8cef563-46df-4aa8-8ad8-8528449cdf97-kube-api-access-64bgl\") pod \"community-operators-wnmzs\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.484256 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64bgl\" (UniqueName: \"kubernetes.io/projected/f8cef563-46df-4aa8-8ad8-8528449cdf97-kube-api-access-64bgl\") pod \"community-operators-wnmzs\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.484441 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-catalog-content\") pod \"community-operators-wnmzs\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.484506 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-utilities\") pod \"community-operators-wnmzs\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.485078 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-catalog-content\") pod \"community-operators-wnmzs\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.485327 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-utilities\") pod \"community-operators-wnmzs\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.512304 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64bgl\" (UniqueName: \"kubernetes.io/projected/f8cef563-46df-4aa8-8ad8-8528449cdf97-kube-api-access-64bgl\") pod \"community-operators-wnmzs\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:09 crc kubenswrapper[4789]: I1008 14:35:09.590486 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:10 crc kubenswrapper[4789]: I1008 14:35:10.091699 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wnmzs"] Oct 08 14:35:10 crc kubenswrapper[4789]: I1008 14:35:10.549904 4789 generic.go:334] "Generic (PLEG): container finished" podID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerID="8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2" exitCode=0 Oct 08 14:35:10 crc kubenswrapper[4789]: I1008 14:35:10.549953 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wnmzs" event={"ID":"f8cef563-46df-4aa8-8ad8-8528449cdf97","Type":"ContainerDied","Data":"8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2"} Oct 08 14:35:10 crc kubenswrapper[4789]: I1008 14:35:10.549981 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wnmzs" event={"ID":"f8cef563-46df-4aa8-8ad8-8528449cdf97","Type":"ContainerStarted","Data":"7b6211b2524645e0eb4fcd30ddcd6f08b95b7be3de26dac6f0119598907eb465"} Oct 08 14:35:11 crc kubenswrapper[4789]: I1008 14:35:11.561305 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wnmzs" event={"ID":"f8cef563-46df-4aa8-8ad8-8528449cdf97","Type":"ContainerStarted","Data":"3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f"} Oct 08 14:35:12 crc kubenswrapper[4789]: I1008 14:35:12.575769 4789 generic.go:334] "Generic (PLEG): container finished" podID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerID="3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f" exitCode=0 Oct 08 14:35:12 crc kubenswrapper[4789]: I1008 14:35:12.575819 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wnmzs" event={"ID":"f8cef563-46df-4aa8-8ad8-8528449cdf97","Type":"ContainerDied","Data":"3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f"} Oct 08 14:35:13 crc kubenswrapper[4789]: I1008 14:35:13.585896 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wnmzs" event={"ID":"f8cef563-46df-4aa8-8ad8-8528449cdf97","Type":"ContainerStarted","Data":"667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94"} Oct 08 14:35:13 crc kubenswrapper[4789]: I1008 14:35:13.606227 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wnmzs" podStartSLOduration=2.139762911 podStartE2EDuration="4.606209961s" podCreationTimestamp="2025-10-08 14:35:09 +0000 UTC" firstStartedPulling="2025-10-08 14:35:10.553369519 +0000 UTC m=+2050.460117001" lastFinishedPulling="2025-10-08 14:35:13.019816569 +0000 UTC m=+2052.926564051" observedRunningTime="2025-10-08 14:35:13.602109418 +0000 UTC m=+2053.508856910" watchObservedRunningTime="2025-10-08 14:35:13.606209961 +0000 UTC m=+2053.512957453" Oct 08 14:35:19 crc kubenswrapper[4789]: I1008 14:35:19.591656 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:19 crc kubenswrapper[4789]: I1008 14:35:19.592334 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:19 crc kubenswrapper[4789]: I1008 14:35:19.649717 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:19 crc kubenswrapper[4789]: I1008 14:35:19.709742 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:19 crc kubenswrapper[4789]: I1008 14:35:19.894328 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wnmzs"] Oct 08 14:35:21 crc kubenswrapper[4789]: I1008 14:35:21.669284 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wnmzs" podUID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerName="registry-server" containerID="cri-o://667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94" gracePeriod=2 Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.150378 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.235326 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-catalog-content\") pod \"f8cef563-46df-4aa8-8ad8-8528449cdf97\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.235402 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64bgl\" (UniqueName: \"kubernetes.io/projected/f8cef563-46df-4aa8-8ad8-8528449cdf97-kube-api-access-64bgl\") pod \"f8cef563-46df-4aa8-8ad8-8528449cdf97\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.235434 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-utilities\") pod \"f8cef563-46df-4aa8-8ad8-8528449cdf97\" (UID: \"f8cef563-46df-4aa8-8ad8-8528449cdf97\") " Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.236453 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-utilities" (OuterVolumeSpecName: "utilities") pod "f8cef563-46df-4aa8-8ad8-8528449cdf97" (UID: "f8cef563-46df-4aa8-8ad8-8528449cdf97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.241127 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8cef563-46df-4aa8-8ad8-8528449cdf97-kube-api-access-64bgl" (OuterVolumeSpecName: "kube-api-access-64bgl") pod "f8cef563-46df-4aa8-8ad8-8528449cdf97" (UID: "f8cef563-46df-4aa8-8ad8-8528449cdf97"). InnerVolumeSpecName "kube-api-access-64bgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.339313 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64bgl\" (UniqueName: \"kubernetes.io/projected/f8cef563-46df-4aa8-8ad8-8528449cdf97-kube-api-access-64bgl\") on node \"crc\" DevicePath \"\"" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.339358 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.689430 4789 generic.go:334] "Generic (PLEG): container finished" podID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerID="667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94" exitCode=0 Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.689516 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wnmzs" event={"ID":"f8cef563-46df-4aa8-8ad8-8528449cdf97","Type":"ContainerDied","Data":"667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94"} Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.690853 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wnmzs" event={"ID":"f8cef563-46df-4aa8-8ad8-8528449cdf97","Type":"ContainerDied","Data":"7b6211b2524645e0eb4fcd30ddcd6f08b95b7be3de26dac6f0119598907eb465"} Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.689586 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wnmzs" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.690895 4789 scope.go:117] "RemoveContainer" containerID="667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.718694 4789 scope.go:117] "RemoveContainer" containerID="3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.767217 4789 scope.go:117] "RemoveContainer" containerID="8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.811444 4789 scope.go:117] "RemoveContainer" containerID="667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94" Oct 08 14:35:22 crc kubenswrapper[4789]: E1008 14:35:22.811846 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94\": container with ID starting with 667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94 not found: ID does not exist" containerID="667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.811903 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94"} err="failed to get container status \"667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94\": rpc error: code = NotFound desc = could not find container \"667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94\": container with ID starting with 667d10868ae46f1687e81c270a9599fd360248d2cc47051de924e5e4282c2b94 not found: ID does not exist" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.811931 4789 scope.go:117] "RemoveContainer" containerID="3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f" Oct 08 14:35:22 crc kubenswrapper[4789]: E1008 14:35:22.812500 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f\": container with ID starting with 3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f not found: ID does not exist" containerID="3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.812543 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f"} err="failed to get container status \"3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f\": rpc error: code = NotFound desc = could not find container \"3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f\": container with ID starting with 3a824936f923f4c1b8d5e3c6f881db1ab0ed6a70975d3e68e40bcd97ab1b148f not found: ID does not exist" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.812570 4789 scope.go:117] "RemoveContainer" containerID="8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2" Oct 08 14:35:22 crc kubenswrapper[4789]: E1008 14:35:22.813152 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2\": container with ID starting with 8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2 not found: ID does not exist" containerID="8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.813220 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2"} err="failed to get container status \"8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2\": rpc error: code = NotFound desc = could not find container \"8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2\": container with ID starting with 8dac8f9bb86ea51559b31d5a94e4d0779eb3186cadadf6d2a48b762443e51cd2 not found: ID does not exist" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.931353 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8cef563-46df-4aa8-8ad8-8528449cdf97" (UID: "f8cef563-46df-4aa8-8ad8-8528449cdf97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.955734 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cef563-46df-4aa8-8ad8-8528449cdf97-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:35:22 crc kubenswrapper[4789]: I1008 14:35:22.999191 4789 scope.go:117] "RemoveContainer" containerID="2f006ed9c2457e36fab03b66b0e890e963638ae6dcc79112e3884b1c14a9d6d7" Oct 08 14:35:23 crc kubenswrapper[4789]: I1008 14:35:23.049044 4789 scope.go:117] "RemoveContainer" containerID="ccbe308e2801588c3627cf00337a6708ad65f29cc86ef71ba89a14e63d4ee497" Oct 08 14:35:23 crc kubenswrapper[4789]: I1008 14:35:23.106028 4789 scope.go:117] "RemoveContainer" containerID="508deddf75c87868a482535af0949a762abe0b0521a1144c6bb3ee5c019f6ca3" Oct 08 14:35:23 crc kubenswrapper[4789]: I1008 14:35:23.172662 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wnmzs"] Oct 08 14:35:23 crc kubenswrapper[4789]: I1008 14:35:23.180446 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wnmzs"] Oct 08 14:35:24 crc kubenswrapper[4789]: I1008 14:35:24.753698 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8cef563-46df-4aa8-8ad8-8528449cdf97" path="/var/lib/kubelet/pods/f8cef563-46df-4aa8-8ad8-8528449cdf97/volumes" Oct 08 14:35:26 crc kubenswrapper[4789]: I1008 14:35:26.432709 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:35:26 crc kubenswrapper[4789]: I1008 14:35:26.433101 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:35:26 crc kubenswrapper[4789]: I1008 14:35:26.433148 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:35:26 crc kubenswrapper[4789]: I1008 14:35:26.433913 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92b1a760457dbd1ac1a6a96f521ea2565954a40aa5a1f3f3fff338d563639227"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:35:26 crc kubenswrapper[4789]: I1008 14:35:26.433977 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://92b1a760457dbd1ac1a6a96f521ea2565954a40aa5a1f3f3fff338d563639227" gracePeriod=600 Oct 08 14:35:26 crc kubenswrapper[4789]: I1008 14:35:26.735245 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="92b1a760457dbd1ac1a6a96f521ea2565954a40aa5a1f3f3fff338d563639227" exitCode=0 Oct 08 14:35:26 crc kubenswrapper[4789]: I1008 14:35:26.743580 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"92b1a760457dbd1ac1a6a96f521ea2565954a40aa5a1f3f3fff338d563639227"} Oct 08 14:35:26 crc kubenswrapper[4789]: I1008 14:35:26.743650 4789 scope.go:117] "RemoveContainer" containerID="7c7cf1a5ddd0fc685a527a55b296390cc3bcc0de9f2433b926b0f274b941ae30" Oct 08 14:35:27 crc kubenswrapper[4789]: I1008 14:35:27.749762 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b"} Oct 08 14:35:28 crc kubenswrapper[4789]: I1008 14:35:28.760355 4789 generic.go:334] "Generic (PLEG): container finished" podID="5a8f80d9-3b33-479c-9d6a-77867520218d" containerID="83da06805f658a5350d690de0396b6f0156385ca9fbc3ba49b23f111bb725aa5" exitCode=0 Oct 08 14:35:28 crc kubenswrapper[4789]: I1008 14:35:28.760453 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" event={"ID":"5a8f80d9-3b33-479c-9d6a-77867520218d","Type":"ContainerDied","Data":"83da06805f658a5350d690de0396b6f0156385ca9fbc3ba49b23f111bb725aa5"} Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.190407 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.300444 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-ssh-key\") pod \"5a8f80d9-3b33-479c-9d6a-77867520218d\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.300656 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbcpn\" (UniqueName: \"kubernetes.io/projected/5a8f80d9-3b33-479c-9d6a-77867520218d-kube-api-access-hbcpn\") pod \"5a8f80d9-3b33-479c-9d6a-77867520218d\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.300725 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-inventory\") pod \"5a8f80d9-3b33-479c-9d6a-77867520218d\" (UID: \"5a8f80d9-3b33-479c-9d6a-77867520218d\") " Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.306193 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a8f80d9-3b33-479c-9d6a-77867520218d-kube-api-access-hbcpn" (OuterVolumeSpecName: "kube-api-access-hbcpn") pod "5a8f80d9-3b33-479c-9d6a-77867520218d" (UID: "5a8f80d9-3b33-479c-9d6a-77867520218d"). InnerVolumeSpecName "kube-api-access-hbcpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.329322 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-inventory" (OuterVolumeSpecName: "inventory") pod "5a8f80d9-3b33-479c-9d6a-77867520218d" (UID: "5a8f80d9-3b33-479c-9d6a-77867520218d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.333567 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5a8f80d9-3b33-479c-9d6a-77867520218d" (UID: "5a8f80d9-3b33-479c-9d6a-77867520218d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.403699 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbcpn\" (UniqueName: \"kubernetes.io/projected/5a8f80d9-3b33-479c-9d6a-77867520218d-kube-api-access-hbcpn\") on node \"crc\" DevicePath \"\"" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.403733 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.403745 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a8f80d9-3b33-479c-9d6a-77867520218d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.778906 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" event={"ID":"5a8f80d9-3b33-479c-9d6a-77867520218d","Type":"ContainerDied","Data":"e9e1117f101ab3902e396fde109d2a78fb234f8393e944acbebcfd92c481b2ce"} Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.779274 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e1117f101ab3902e396fde109d2a78fb234f8393e944acbebcfd92c481b2ce" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.778978 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xpffg" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.860902 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w"] Oct 08 14:35:30 crc kubenswrapper[4789]: E1008 14:35:30.861299 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerName="extract-content" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.861324 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerName="extract-content" Oct 08 14:35:30 crc kubenswrapper[4789]: E1008 14:35:30.861349 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerName="extract-utilities" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.861356 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerName="extract-utilities" Oct 08 14:35:30 crc kubenswrapper[4789]: E1008 14:35:30.861376 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerName="registry-server" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.861382 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerName="registry-server" Oct 08 14:35:30 crc kubenswrapper[4789]: E1008 14:35:30.861393 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a8f80d9-3b33-479c-9d6a-77867520218d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.861400 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a8f80d9-3b33-479c-9d6a-77867520218d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.861598 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8cef563-46df-4aa8-8ad8-8528449cdf97" containerName="registry-server" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.861629 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a8f80d9-3b33-479c-9d6a-77867520218d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.862271 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.865637 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.865752 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.865637 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.866719 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.873327 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w"] Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.913669 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-znv9w\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.913946 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-znv9w\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:30 crc kubenswrapper[4789]: I1008 14:35:30.914067 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dp5f\" (UniqueName: \"kubernetes.io/projected/167cf260-5310-4c48-9911-b5257c1ba256-kube-api-access-2dp5f\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-znv9w\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.015685 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-znv9w\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.016075 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-znv9w\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.016106 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dp5f\" (UniqueName: \"kubernetes.io/projected/167cf260-5310-4c48-9911-b5257c1ba256-kube-api-access-2dp5f\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-znv9w\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.021498 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-znv9w\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.023486 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-znv9w\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.042616 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dp5f\" (UniqueName: \"kubernetes.io/projected/167cf260-5310-4c48-9911-b5257c1ba256-kube-api-access-2dp5f\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-znv9w\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.043485 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-4x2xn"] Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.052143 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-4x2xn"] Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.183456 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.689176 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w"] Oct 08 14:35:31 crc kubenswrapper[4789]: I1008 14:35:31.788193 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" event={"ID":"167cf260-5310-4c48-9911-b5257c1ba256","Type":"ContainerStarted","Data":"d095ce9d28e4136972b18313a1126edbb47646cfc66276156769c7f261c63e41"} Oct 08 14:35:32 crc kubenswrapper[4789]: I1008 14:35:32.741375 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8607d7e0-182c-4dc0-b8a6-2baeb594e61f" path="/var/lib/kubelet/pods/8607d7e0-182c-4dc0-b8a6-2baeb594e61f/volumes" Oct 08 14:35:32 crc kubenswrapper[4789]: I1008 14:35:32.800682 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" event={"ID":"167cf260-5310-4c48-9911-b5257c1ba256","Type":"ContainerStarted","Data":"6e6e5b197b75135686132e5267b846170548815e1014b26d902b7ad22234b966"} Oct 08 14:35:32 crc kubenswrapper[4789]: I1008 14:35:32.819390 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" podStartSLOduration=2.405883775 podStartE2EDuration="2.819367173s" podCreationTimestamp="2025-10-08 14:35:30 +0000 UTC" firstStartedPulling="2025-10-08 14:35:31.695029224 +0000 UTC m=+2071.601776716" lastFinishedPulling="2025-10-08 14:35:32.108512622 +0000 UTC m=+2072.015260114" observedRunningTime="2025-10-08 14:35:32.816273609 +0000 UTC m=+2072.723021101" watchObservedRunningTime="2025-10-08 14:35:32.819367173 +0000 UTC m=+2072.726114665" Oct 08 14:36:23 crc kubenswrapper[4789]: I1008 14:36:23.275582 4789 scope.go:117] "RemoveContainer" containerID="62da5d1fa3ec95c2034bc976c16573164d1fbeb0f80e110ec25680f4ae6b2ba1" Oct 08 14:36:32 crc kubenswrapper[4789]: I1008 14:36:32.325422 4789 generic.go:334] "Generic (PLEG): container finished" podID="167cf260-5310-4c48-9911-b5257c1ba256" containerID="6e6e5b197b75135686132e5267b846170548815e1014b26d902b7ad22234b966" exitCode=2 Oct 08 14:36:32 crc kubenswrapper[4789]: I1008 14:36:32.325543 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" event={"ID":"167cf260-5310-4c48-9911-b5257c1ba256","Type":"ContainerDied","Data":"6e6e5b197b75135686132e5267b846170548815e1014b26d902b7ad22234b966"} Oct 08 14:36:33 crc kubenswrapper[4789]: I1008 14:36:33.818729 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:36:33 crc kubenswrapper[4789]: I1008 14:36:33.953219 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-inventory\") pod \"167cf260-5310-4c48-9911-b5257c1ba256\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " Oct 08 14:36:33 crc kubenswrapper[4789]: I1008 14:36:33.953266 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-ssh-key\") pod \"167cf260-5310-4c48-9911-b5257c1ba256\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " Oct 08 14:36:33 crc kubenswrapper[4789]: I1008 14:36:33.953349 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dp5f\" (UniqueName: \"kubernetes.io/projected/167cf260-5310-4c48-9911-b5257c1ba256-kube-api-access-2dp5f\") pod \"167cf260-5310-4c48-9911-b5257c1ba256\" (UID: \"167cf260-5310-4c48-9911-b5257c1ba256\") " Oct 08 14:36:33 crc kubenswrapper[4789]: I1008 14:36:33.965387 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/167cf260-5310-4c48-9911-b5257c1ba256-kube-api-access-2dp5f" (OuterVolumeSpecName: "kube-api-access-2dp5f") pod "167cf260-5310-4c48-9911-b5257c1ba256" (UID: "167cf260-5310-4c48-9911-b5257c1ba256"). InnerVolumeSpecName "kube-api-access-2dp5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:36:33 crc kubenswrapper[4789]: I1008 14:36:33.982110 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-inventory" (OuterVolumeSpecName: "inventory") pod "167cf260-5310-4c48-9911-b5257c1ba256" (UID: "167cf260-5310-4c48-9911-b5257c1ba256"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:36:34 crc kubenswrapper[4789]: I1008 14:36:34.010540 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "167cf260-5310-4c48-9911-b5257c1ba256" (UID: "167cf260-5310-4c48-9911-b5257c1ba256"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:36:34 crc kubenswrapper[4789]: I1008 14:36:34.055772 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dp5f\" (UniqueName: \"kubernetes.io/projected/167cf260-5310-4c48-9911-b5257c1ba256-kube-api-access-2dp5f\") on node \"crc\" DevicePath \"\"" Oct 08 14:36:34 crc kubenswrapper[4789]: I1008 14:36:34.055872 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:36:34 crc kubenswrapper[4789]: I1008 14:36:34.055889 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/167cf260-5310-4c48-9911-b5257c1ba256-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:36:34 crc kubenswrapper[4789]: I1008 14:36:34.343329 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" event={"ID":"167cf260-5310-4c48-9911-b5257c1ba256","Type":"ContainerDied","Data":"d095ce9d28e4136972b18313a1126edbb47646cfc66276156769c7f261c63e41"} Oct 08 14:36:34 crc kubenswrapper[4789]: I1008 14:36:34.343396 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-znv9w" Oct 08 14:36:34 crc kubenswrapper[4789]: I1008 14:36:34.343402 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d095ce9d28e4136972b18313a1126edbb47646cfc66276156769c7f261c63e41" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.034002 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd"] Oct 08 14:36:42 crc kubenswrapper[4789]: E1008 14:36:42.034853 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="167cf260-5310-4c48-9911-b5257c1ba256" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.034867 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="167cf260-5310-4c48-9911-b5257c1ba256" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.035068 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="167cf260-5310-4c48-9911-b5257c1ba256" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.035713 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.037756 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.038745 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.038996 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.044578 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.088724 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd"] Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.109073 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.109136 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcp6v\" (UniqueName: \"kubernetes.io/projected/6f2ef14f-0023-4c8d-8153-5dba30e84c17-kube-api-access-fcp6v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.109270 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.215007 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.215082 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcp6v\" (UniqueName: \"kubernetes.io/projected/6f2ef14f-0023-4c8d-8153-5dba30e84c17-kube-api-access-fcp6v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.215239 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.231251 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.238616 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.246433 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcp6v\" (UniqueName: \"kubernetes.io/projected/6f2ef14f-0023-4c8d-8153-5dba30e84c17-kube-api-access-fcp6v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.366200 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:36:42 crc kubenswrapper[4789]: I1008 14:36:42.934438 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd"] Oct 08 14:36:42 crc kubenswrapper[4789]: W1008 14:36:42.938196 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f2ef14f_0023_4c8d_8153_5dba30e84c17.slice/crio-c8dae8a5b62ca16d97d3e9a0552995ca504e9ec3750a591565bb5957d5d28963 WatchSource:0}: Error finding container c8dae8a5b62ca16d97d3e9a0552995ca504e9ec3750a591565bb5957d5d28963: Status 404 returned error can't find the container with id c8dae8a5b62ca16d97d3e9a0552995ca504e9ec3750a591565bb5957d5d28963 Oct 08 14:36:43 crc kubenswrapper[4789]: I1008 14:36:43.436575 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" event={"ID":"6f2ef14f-0023-4c8d-8153-5dba30e84c17","Type":"ContainerStarted","Data":"c8dae8a5b62ca16d97d3e9a0552995ca504e9ec3750a591565bb5957d5d28963"} Oct 08 14:36:44 crc kubenswrapper[4789]: I1008 14:36:44.462287 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" event={"ID":"6f2ef14f-0023-4c8d-8153-5dba30e84c17","Type":"ContainerStarted","Data":"bc1f129ecb2d28906c0ce72cbf9b4e06406e50f66bd19d17d7aa71bfa6744615"} Oct 08 14:36:44 crc kubenswrapper[4789]: I1008 14:36:44.492198 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" podStartSLOduration=1.961089348 podStartE2EDuration="2.492163244s" podCreationTimestamp="2025-10-08 14:36:42 +0000 UTC" firstStartedPulling="2025-10-08 14:36:42.940338673 +0000 UTC m=+2142.847086175" lastFinishedPulling="2025-10-08 14:36:43.471412579 +0000 UTC m=+2143.378160071" observedRunningTime="2025-10-08 14:36:44.478321457 +0000 UTC m=+2144.385068969" watchObservedRunningTime="2025-10-08 14:36:44.492163244 +0000 UTC m=+2144.398910736" Oct 08 14:37:26 crc kubenswrapper[4789]: I1008 14:37:26.433267 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:37:26 crc kubenswrapper[4789]: I1008 14:37:26.433887 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:37:34 crc kubenswrapper[4789]: I1008 14:37:34.956584 4789 generic.go:334] "Generic (PLEG): container finished" podID="6f2ef14f-0023-4c8d-8153-5dba30e84c17" containerID="bc1f129ecb2d28906c0ce72cbf9b4e06406e50f66bd19d17d7aa71bfa6744615" exitCode=0 Oct 08 14:37:34 crc kubenswrapper[4789]: I1008 14:37:34.956672 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" event={"ID":"6f2ef14f-0023-4c8d-8153-5dba30e84c17","Type":"ContainerDied","Data":"bc1f129ecb2d28906c0ce72cbf9b4e06406e50f66bd19d17d7aa71bfa6744615"} Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.399960 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.489600 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcp6v\" (UniqueName: \"kubernetes.io/projected/6f2ef14f-0023-4c8d-8153-5dba30e84c17-kube-api-access-fcp6v\") pod \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.489724 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-ssh-key\") pod \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.489749 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-inventory\") pod \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\" (UID: \"6f2ef14f-0023-4c8d-8153-5dba30e84c17\") " Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.495441 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f2ef14f-0023-4c8d-8153-5dba30e84c17-kube-api-access-fcp6v" (OuterVolumeSpecName: "kube-api-access-fcp6v") pod "6f2ef14f-0023-4c8d-8153-5dba30e84c17" (UID: "6f2ef14f-0023-4c8d-8153-5dba30e84c17"). InnerVolumeSpecName "kube-api-access-fcp6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.518002 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-inventory" (OuterVolumeSpecName: "inventory") pod "6f2ef14f-0023-4c8d-8153-5dba30e84c17" (UID: "6f2ef14f-0023-4c8d-8153-5dba30e84c17"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.523433 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6f2ef14f-0023-4c8d-8153-5dba30e84c17" (UID: "6f2ef14f-0023-4c8d-8153-5dba30e84c17"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.593109 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.593163 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f2ef14f-0023-4c8d-8153-5dba30e84c17-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.593183 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcp6v\" (UniqueName: \"kubernetes.io/projected/6f2ef14f-0023-4c8d-8153-5dba30e84c17-kube-api-access-fcp6v\") on node \"crc\" DevicePath \"\"" Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.996967 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" event={"ID":"6f2ef14f-0023-4c8d-8153-5dba30e84c17","Type":"ContainerDied","Data":"c8dae8a5b62ca16d97d3e9a0552995ca504e9ec3750a591565bb5957d5d28963"} Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.997022 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8dae8a5b62ca16d97d3e9a0552995ca504e9ec3750a591565bb5957d5d28963" Oct 08 14:37:36 crc kubenswrapper[4789]: I1008 14:37:36.997126 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.090359 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-29nmw"] Oct 08 14:37:37 crc kubenswrapper[4789]: E1008 14:37:37.090754 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f2ef14f-0023-4c8d-8153-5dba30e84c17" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.090767 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f2ef14f-0023-4c8d-8153-5dba30e84c17" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.091099 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f2ef14f-0023-4c8d-8153-5dba30e84c17" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.091774 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.093845 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.094112 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.094293 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.096098 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.100890 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-29nmw"] Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.106238 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-29nmw\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.106373 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-29nmw\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.106400 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27wcx\" (UniqueName: \"kubernetes.io/projected/ad2db88c-41b1-4feb-b579-96bb456cea2a-kube-api-access-27wcx\") pod \"ssh-known-hosts-edpm-deployment-29nmw\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.208524 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-29nmw\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.208828 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27wcx\" (UniqueName: \"kubernetes.io/projected/ad2db88c-41b1-4feb-b579-96bb456cea2a-kube-api-access-27wcx\") pod \"ssh-known-hosts-edpm-deployment-29nmw\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.208902 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-29nmw\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.215491 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-29nmw\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.224332 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-29nmw\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.241981 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27wcx\" (UniqueName: \"kubernetes.io/projected/ad2db88c-41b1-4feb-b579-96bb456cea2a-kube-api-access-27wcx\") pod \"ssh-known-hosts-edpm-deployment-29nmw\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.422440 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:37 crc kubenswrapper[4789]: I1008 14:37:37.905213 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-29nmw"] Oct 08 14:37:38 crc kubenswrapper[4789]: I1008 14:37:38.007892 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" event={"ID":"ad2db88c-41b1-4feb-b579-96bb456cea2a","Type":"ContainerStarted","Data":"eb256a11bae9df305977c20091360b7f4c6af717634d13b7bf848b722a17f52e"} Oct 08 14:37:39 crc kubenswrapper[4789]: I1008 14:37:39.016823 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" event={"ID":"ad2db88c-41b1-4feb-b579-96bb456cea2a","Type":"ContainerStarted","Data":"8a8292fb21bf49665345abcc2878b840238d09004a4fad90493d85b90489b9d0"} Oct 08 14:37:39 crc kubenswrapper[4789]: I1008 14:37:39.039483 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" podStartSLOduration=1.579383942 podStartE2EDuration="2.039460616s" podCreationTimestamp="2025-10-08 14:37:37 +0000 UTC" firstStartedPulling="2025-10-08 14:37:37.914145057 +0000 UTC m=+2197.820892559" lastFinishedPulling="2025-10-08 14:37:38.374221741 +0000 UTC m=+2198.280969233" observedRunningTime="2025-10-08 14:37:39.032512847 +0000 UTC m=+2198.939260339" watchObservedRunningTime="2025-10-08 14:37:39.039460616 +0000 UTC m=+2198.946208108" Oct 08 14:37:46 crc kubenswrapper[4789]: I1008 14:37:46.078867 4789 generic.go:334] "Generic (PLEG): container finished" podID="ad2db88c-41b1-4feb-b579-96bb456cea2a" containerID="8a8292fb21bf49665345abcc2878b840238d09004a4fad90493d85b90489b9d0" exitCode=0 Oct 08 14:37:46 crc kubenswrapper[4789]: I1008 14:37:46.078962 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" event={"ID":"ad2db88c-41b1-4feb-b579-96bb456cea2a","Type":"ContainerDied","Data":"8a8292fb21bf49665345abcc2878b840238d09004a4fad90493d85b90489b9d0"} Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.504816 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.635786 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-ssh-key-openstack-edpm-ipam\") pod \"ad2db88c-41b1-4feb-b579-96bb456cea2a\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.635884 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27wcx\" (UniqueName: \"kubernetes.io/projected/ad2db88c-41b1-4feb-b579-96bb456cea2a-kube-api-access-27wcx\") pod \"ad2db88c-41b1-4feb-b579-96bb456cea2a\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.635901 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-inventory-0\") pod \"ad2db88c-41b1-4feb-b579-96bb456cea2a\" (UID: \"ad2db88c-41b1-4feb-b579-96bb456cea2a\") " Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.641134 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad2db88c-41b1-4feb-b579-96bb456cea2a-kube-api-access-27wcx" (OuterVolumeSpecName: "kube-api-access-27wcx") pod "ad2db88c-41b1-4feb-b579-96bb456cea2a" (UID: "ad2db88c-41b1-4feb-b579-96bb456cea2a"). InnerVolumeSpecName "kube-api-access-27wcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.661472 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ad2db88c-41b1-4feb-b579-96bb456cea2a" (UID: "ad2db88c-41b1-4feb-b579-96bb456cea2a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.682283 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "ad2db88c-41b1-4feb-b579-96bb456cea2a" (UID: "ad2db88c-41b1-4feb-b579-96bb456cea2a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.738591 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27wcx\" (UniqueName: \"kubernetes.io/projected/ad2db88c-41b1-4feb-b579-96bb456cea2a-kube-api-access-27wcx\") on node \"crc\" DevicePath \"\"" Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.738622 4789 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:37:47 crc kubenswrapper[4789]: I1008 14:37:47.738635 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad2db88c-41b1-4feb-b579-96bb456cea2a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.097785 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" event={"ID":"ad2db88c-41b1-4feb-b579-96bb456cea2a","Type":"ContainerDied","Data":"eb256a11bae9df305977c20091360b7f4c6af717634d13b7bf848b722a17f52e"} Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.098184 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb256a11bae9df305977c20091360b7f4c6af717634d13b7bf848b722a17f52e" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.097859 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-29nmw" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.185318 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th"] Oct 08 14:37:48 crc kubenswrapper[4789]: E1008 14:37:48.185793 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad2db88c-41b1-4feb-b579-96bb456cea2a" containerName="ssh-known-hosts-edpm-deployment" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.185812 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad2db88c-41b1-4feb-b579-96bb456cea2a" containerName="ssh-known-hosts-edpm-deployment" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.186078 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad2db88c-41b1-4feb-b579-96bb456cea2a" containerName="ssh-known-hosts-edpm-deployment" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.186789 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.194658 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.194959 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.194662 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.195594 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.217635 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th"] Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.350598 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xz5th\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.350692 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xz5th\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.350717 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgcd2\" (UniqueName: \"kubernetes.io/projected/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-kube-api-access-jgcd2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xz5th\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.452463 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xz5th\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.452537 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xz5th\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.452591 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgcd2\" (UniqueName: \"kubernetes.io/projected/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-kube-api-access-jgcd2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xz5th\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.458722 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xz5th\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.459088 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xz5th\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.469512 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgcd2\" (UniqueName: \"kubernetes.io/projected/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-kube-api-access-jgcd2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xz5th\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:48 crc kubenswrapper[4789]: I1008 14:37:48.514804 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:49 crc kubenswrapper[4789]: I1008 14:37:49.046316 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th"] Oct 08 14:37:49 crc kubenswrapper[4789]: I1008 14:37:49.106558 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" event={"ID":"1ef78bfb-47eb-40df-93d6-7b6070c33f5f","Type":"ContainerStarted","Data":"b329ab8448605384c4aa94c26d263a593d888c30671324c95c52c0b5a7f4aee1"} Oct 08 14:37:50 crc kubenswrapper[4789]: I1008 14:37:50.117128 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" event={"ID":"1ef78bfb-47eb-40df-93d6-7b6070c33f5f","Type":"ContainerStarted","Data":"e5d36182c669db6237dd4f04a62b695f3dee3da8228b6cfd94151ba3423e7645"} Oct 08 14:37:50 crc kubenswrapper[4789]: I1008 14:37:50.146727 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" podStartSLOduration=1.721560661 podStartE2EDuration="2.146706266s" podCreationTimestamp="2025-10-08 14:37:48 +0000 UTC" firstStartedPulling="2025-10-08 14:37:49.049629304 +0000 UTC m=+2208.956376796" lastFinishedPulling="2025-10-08 14:37:49.474774889 +0000 UTC m=+2209.381522401" observedRunningTime="2025-10-08 14:37:50.140096266 +0000 UTC m=+2210.046843798" watchObservedRunningTime="2025-10-08 14:37:50.146706266 +0000 UTC m=+2210.053453758" Oct 08 14:37:56 crc kubenswrapper[4789]: I1008 14:37:56.432913 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:37:56 crc kubenswrapper[4789]: I1008 14:37:56.433471 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:37:58 crc kubenswrapper[4789]: I1008 14:37:58.201118 4789 generic.go:334] "Generic (PLEG): container finished" podID="1ef78bfb-47eb-40df-93d6-7b6070c33f5f" containerID="e5d36182c669db6237dd4f04a62b695f3dee3da8228b6cfd94151ba3423e7645" exitCode=0 Oct 08 14:37:58 crc kubenswrapper[4789]: I1008 14:37:58.201203 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" event={"ID":"1ef78bfb-47eb-40df-93d6-7b6070c33f5f","Type":"ContainerDied","Data":"e5d36182c669db6237dd4f04a62b695f3dee3da8228b6cfd94151ba3423e7645"} Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.657352 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.734773 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-ssh-key\") pod \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.734892 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgcd2\" (UniqueName: \"kubernetes.io/projected/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-kube-api-access-jgcd2\") pod \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.734962 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-inventory\") pod \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\" (UID: \"1ef78bfb-47eb-40df-93d6-7b6070c33f5f\") " Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.740906 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-kube-api-access-jgcd2" (OuterVolumeSpecName: "kube-api-access-jgcd2") pod "1ef78bfb-47eb-40df-93d6-7b6070c33f5f" (UID: "1ef78bfb-47eb-40df-93d6-7b6070c33f5f"). InnerVolumeSpecName "kube-api-access-jgcd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.761805 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-inventory" (OuterVolumeSpecName: "inventory") pod "1ef78bfb-47eb-40df-93d6-7b6070c33f5f" (UID: "1ef78bfb-47eb-40df-93d6-7b6070c33f5f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.764282 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1ef78bfb-47eb-40df-93d6-7b6070c33f5f" (UID: "1ef78bfb-47eb-40df-93d6-7b6070c33f5f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.837426 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.837457 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgcd2\" (UniqueName: \"kubernetes.io/projected/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-kube-api-access-jgcd2\") on node \"crc\" DevicePath \"\"" Oct 08 14:37:59 crc kubenswrapper[4789]: I1008 14:37:59.837471 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef78bfb-47eb-40df-93d6-7b6070c33f5f-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.229775 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" event={"ID":"1ef78bfb-47eb-40df-93d6-7b6070c33f5f","Type":"ContainerDied","Data":"b329ab8448605384c4aa94c26d263a593d888c30671324c95c52c0b5a7f4aee1"} Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.229831 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xz5th" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.229842 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b329ab8448605384c4aa94c26d263a593d888c30671324c95c52c0b5a7f4aee1" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.305743 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg"] Oct 08 14:38:00 crc kubenswrapper[4789]: E1008 14:38:00.306267 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef78bfb-47eb-40df-93d6-7b6070c33f5f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.306293 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef78bfb-47eb-40df-93d6-7b6070c33f5f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.306527 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef78bfb-47eb-40df-93d6-7b6070c33f5f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.307242 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.309007 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.309251 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.309430 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.310579 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.313827 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg"] Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.358102 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.358430 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.358466 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq5q7\" (UniqueName: \"kubernetes.io/projected/f2575ace-ab3d-475e-b049-a2d2614104bf-kube-api-access-bq5q7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.460481 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.460553 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq5q7\" (UniqueName: \"kubernetes.io/projected/f2575ace-ab3d-475e-b049-a2d2614104bf-kube-api-access-bq5q7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.460734 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.465026 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.472267 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.486681 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq5q7\" (UniqueName: \"kubernetes.io/projected/f2575ace-ab3d-475e-b049-a2d2614104bf-kube-api-access-bq5q7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:00 crc kubenswrapper[4789]: I1008 14:38:00.661607 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:01 crc kubenswrapper[4789]: I1008 14:38:01.164741 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg"] Oct 08 14:38:01 crc kubenswrapper[4789]: I1008 14:38:01.241533 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" event={"ID":"f2575ace-ab3d-475e-b049-a2d2614104bf","Type":"ContainerStarted","Data":"74823d1ce2a4dddc6cb346edeadc7cfcf6d5ce1a436153be2b4d375a36d5d4a4"} Oct 08 14:38:01 crc kubenswrapper[4789]: I1008 14:38:01.609042 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:38:02 crc kubenswrapper[4789]: I1008 14:38:02.250911 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" event={"ID":"f2575ace-ab3d-475e-b049-a2d2614104bf","Type":"ContainerStarted","Data":"87dd80c1dfd00d280704560ff857c99bce1cfea7c885aa34a85ef2eaf9d93307"} Oct 08 14:38:02 crc kubenswrapper[4789]: I1008 14:38:02.272166 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" podStartSLOduration=1.844430169 podStartE2EDuration="2.272147293s" podCreationTimestamp="2025-10-08 14:38:00 +0000 UTC" firstStartedPulling="2025-10-08 14:38:01.178841874 +0000 UTC m=+2221.085589366" lastFinishedPulling="2025-10-08 14:38:01.606558998 +0000 UTC m=+2221.513306490" observedRunningTime="2025-10-08 14:38:02.266263363 +0000 UTC m=+2222.173010855" watchObservedRunningTime="2025-10-08 14:38:02.272147293 +0000 UTC m=+2222.178894785" Oct 08 14:38:12 crc kubenswrapper[4789]: I1008 14:38:12.333238 4789 generic.go:334] "Generic (PLEG): container finished" podID="f2575ace-ab3d-475e-b049-a2d2614104bf" containerID="87dd80c1dfd00d280704560ff857c99bce1cfea7c885aa34a85ef2eaf9d93307" exitCode=0 Oct 08 14:38:12 crc kubenswrapper[4789]: I1008 14:38:12.333327 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" event={"ID":"f2575ace-ab3d-475e-b049-a2d2614104bf","Type":"ContainerDied","Data":"87dd80c1dfd00d280704560ff857c99bce1cfea7c885aa34a85ef2eaf9d93307"} Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.748662 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.821271 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-ssh-key\") pod \"f2575ace-ab3d-475e-b049-a2d2614104bf\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.821439 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq5q7\" (UniqueName: \"kubernetes.io/projected/f2575ace-ab3d-475e-b049-a2d2614104bf-kube-api-access-bq5q7\") pod \"f2575ace-ab3d-475e-b049-a2d2614104bf\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.821504 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-inventory\") pod \"f2575ace-ab3d-475e-b049-a2d2614104bf\" (UID: \"f2575ace-ab3d-475e-b049-a2d2614104bf\") " Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.827812 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2575ace-ab3d-475e-b049-a2d2614104bf-kube-api-access-bq5q7" (OuterVolumeSpecName: "kube-api-access-bq5q7") pod "f2575ace-ab3d-475e-b049-a2d2614104bf" (UID: "f2575ace-ab3d-475e-b049-a2d2614104bf"). InnerVolumeSpecName "kube-api-access-bq5q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.854791 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-inventory" (OuterVolumeSpecName: "inventory") pod "f2575ace-ab3d-475e-b049-a2d2614104bf" (UID: "f2575ace-ab3d-475e-b049-a2d2614104bf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.870934 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f2575ace-ab3d-475e-b049-a2d2614104bf" (UID: "f2575ace-ab3d-475e-b049-a2d2614104bf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.922685 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.922729 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f2575ace-ab3d-475e-b049-a2d2614104bf-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:13 crc kubenswrapper[4789]: I1008 14:38:13.922741 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq5q7\" (UniqueName: \"kubernetes.io/projected/f2575ace-ab3d-475e-b049-a2d2614104bf-kube-api-access-bq5q7\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.350248 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" event={"ID":"f2575ace-ab3d-475e-b049-a2d2614104bf","Type":"ContainerDied","Data":"74823d1ce2a4dddc6cb346edeadc7cfcf6d5ce1a436153be2b4d375a36d5d4a4"} Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.350300 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74823d1ce2a4dddc6cb346edeadc7cfcf6d5ce1a436153be2b4d375a36d5d4a4" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.350295 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.512740 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm"] Oct 08 14:38:14 crc kubenswrapper[4789]: E1008 14:38:14.513503 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2575ace-ab3d-475e-b049-a2d2614104bf" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.513528 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2575ace-ab3d-475e-b049-a2d2614104bf" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.513838 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2575ace-ab3d-475e-b049-a2d2614104bf" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.514672 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.519853 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.519874 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.519965 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.520211 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.520237 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.520293 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.520328 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.520346 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.523606 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm"] Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534331 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534480 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534519 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534547 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534591 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534638 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534680 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534721 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534748 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxj4h\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-kube-api-access-cxj4h\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534786 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534819 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534928 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.534961 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.535026 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.636746 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.636877 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.636912 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.636940 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.636963 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.637006 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.637043 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.637078 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.637102 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxj4h\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-kube-api-access-cxj4h\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.637130 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.637167 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.637240 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.637273 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.637320 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.641877 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.641958 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.642219 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.642246 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.642513 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.642563 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.643034 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.643887 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.644223 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.644342 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.644676 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.647399 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.651134 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.658642 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxj4h\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-kube-api-access-cxj4h\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:14 crc kubenswrapper[4789]: I1008 14:38:14.834128 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:15 crc kubenswrapper[4789]: I1008 14:38:15.368032 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm"] Oct 08 14:38:15 crc kubenswrapper[4789]: I1008 14:38:15.376302 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 14:38:16 crc kubenswrapper[4789]: I1008 14:38:16.368489 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" event={"ID":"92d5b86d-9984-45c5-8287-7b3094b4332f","Type":"ContainerStarted","Data":"8a97a975eb5d068e76c4aa7cfeceb55d38a6364356de04f8865dfddcd8ef174e"} Oct 08 14:38:16 crc kubenswrapper[4789]: I1008 14:38:16.368790 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" event={"ID":"92d5b86d-9984-45c5-8287-7b3094b4332f","Type":"ContainerStarted","Data":"7a856a94173a6f4648f42d2ebe5fc879d010f66c67c188e44fd0ff4803810d2c"} Oct 08 14:38:16 crc kubenswrapper[4789]: I1008 14:38:16.389746 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" podStartSLOduration=1.7867174559999999 podStartE2EDuration="2.389727279s" podCreationTimestamp="2025-10-08 14:38:14 +0000 UTC" firstStartedPulling="2025-10-08 14:38:15.376031465 +0000 UTC m=+2235.282778967" lastFinishedPulling="2025-10-08 14:38:15.979041298 +0000 UTC m=+2235.885788790" observedRunningTime="2025-10-08 14:38:16.38535331 +0000 UTC m=+2236.292100822" watchObservedRunningTime="2025-10-08 14:38:16.389727279 +0000 UTC m=+2236.296474771" Oct 08 14:38:26 crc kubenswrapper[4789]: I1008 14:38:26.433178 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:38:26 crc kubenswrapper[4789]: I1008 14:38:26.433777 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:38:26 crc kubenswrapper[4789]: I1008 14:38:26.433834 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:38:26 crc kubenswrapper[4789]: I1008 14:38:26.434606 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:38:26 crc kubenswrapper[4789]: I1008 14:38:26.434653 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" gracePeriod=600 Oct 08 14:38:26 crc kubenswrapper[4789]: E1008 14:38:26.552606 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:38:27 crc kubenswrapper[4789]: I1008 14:38:27.482410 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" exitCode=0 Oct 08 14:38:27 crc kubenswrapper[4789]: I1008 14:38:27.482454 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b"} Oct 08 14:38:27 crc kubenswrapper[4789]: I1008 14:38:27.482809 4789 scope.go:117] "RemoveContainer" containerID="92b1a760457dbd1ac1a6a96f521ea2565954a40aa5a1f3f3fff338d563639227" Oct 08 14:38:27 crc kubenswrapper[4789]: I1008 14:38:27.483899 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:38:27 crc kubenswrapper[4789]: E1008 14:38:27.484194 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:38:40 crc kubenswrapper[4789]: I1008 14:38:40.735871 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:38:40 crc kubenswrapper[4789]: E1008 14:38:40.736497 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:38:54 crc kubenswrapper[4789]: I1008 14:38:54.730859 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:38:54 crc kubenswrapper[4789]: E1008 14:38:54.731719 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:38:55 crc kubenswrapper[4789]: I1008 14:38:55.722808 4789 generic.go:334] "Generic (PLEG): container finished" podID="92d5b86d-9984-45c5-8287-7b3094b4332f" containerID="8a97a975eb5d068e76c4aa7cfeceb55d38a6364356de04f8865dfddcd8ef174e" exitCode=0 Oct 08 14:38:55 crc kubenswrapper[4789]: I1008 14:38:55.722898 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" event={"ID":"92d5b86d-9984-45c5-8287-7b3094b4332f","Type":"ContainerDied","Data":"8a97a975eb5d068e76c4aa7cfeceb55d38a6364356de04f8865dfddcd8ef174e"} Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.231278 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260364 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-nova-combined-ca-bundle\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260433 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-bootstrap-combined-ca-bundle\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260455 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ovn-combined-ca-bundle\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260479 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-telemetry-combined-ca-bundle\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260498 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260538 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-neutron-metadata-combined-ca-bundle\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260569 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxj4h\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-kube-api-access-cxj4h\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260652 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-inventory\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260710 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-repo-setup-combined-ca-bundle\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260769 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ssh-key\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260794 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260838 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.260973 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.261105 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-libvirt-combined-ca-bundle\") pod \"92d5b86d-9984-45c5-8287-7b3094b4332f\" (UID: \"92d5b86d-9984-45c5-8287-7b3094b4332f\") " Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.273079 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.275910 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.280343 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.280634 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.281317 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.281333 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.281356 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.281403 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.281335 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-kube-api-access-cxj4h" (OuterVolumeSpecName: "kube-api-access-cxj4h") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "kube-api-access-cxj4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.281608 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.281787 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.281948 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.309038 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-inventory" (OuterVolumeSpecName: "inventory") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.324326 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "92d5b86d-9984-45c5-8287-7b3094b4332f" (UID: "92d5b86d-9984-45c5-8287-7b3094b4332f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363239 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363275 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363287 4789 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363318 4789 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363328 4789 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363664 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363680 4789 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363690 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363702 4789 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363711 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxj4h\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-kube-api-access-cxj4h\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363748 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363756 4789 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363765 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92d5b86d-9984-45c5-8287-7b3094b4332f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.363773 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/92d5b86d-9984-45c5-8287-7b3094b4332f-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.740314 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" event={"ID":"92d5b86d-9984-45c5-8287-7b3094b4332f","Type":"ContainerDied","Data":"7a856a94173a6f4648f42d2ebe5fc879d010f66c67c188e44fd0ff4803810d2c"} Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.740362 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a856a94173a6f4648f42d2ebe5fc879d010f66c67c188e44fd0ff4803810d2c" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.740378 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.849324 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2"] Oct 08 14:38:57 crc kubenswrapper[4789]: E1008 14:38:57.849778 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d5b86d-9984-45c5-8287-7b3094b4332f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.849794 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d5b86d-9984-45c5-8287-7b3094b4332f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.850043 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d5b86d-9984-45c5-8287-7b3094b4332f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.850748 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.852892 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.853030 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.853104 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.856797 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.858307 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.858950 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2"] Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.880282 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.880378 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.880442 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4t87\" (UniqueName: \"kubernetes.io/projected/516d476a-1a73-4e62-a06f-d9eb90541500-kube-api-access-n4t87\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.880475 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/516d476a-1a73-4e62-a06f-d9eb90541500-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.880548 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.948084 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s5hlk"] Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.950592 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.958536 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s5hlk"] Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.982456 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.982543 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-utilities\") pod \"certified-operators-s5hlk\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.982590 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4t87\" (UniqueName: \"kubernetes.io/projected/516d476a-1a73-4e62-a06f-d9eb90541500-kube-api-access-n4t87\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.982626 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/516d476a-1a73-4e62-a06f-d9eb90541500-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.982668 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2l9p\" (UniqueName: \"kubernetes.io/projected/2286403a-5095-458a-9f73-83b378451237-kube-api-access-j2l9p\") pod \"certified-operators-s5hlk\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.982709 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-catalog-content\") pod \"certified-operators-s5hlk\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.982731 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.982863 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.984239 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/516d476a-1a73-4e62-a06f-d9eb90541500-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.986641 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.986934 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:57 crc kubenswrapper[4789]: I1008 14:38:57.999538 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.006885 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4t87\" (UniqueName: \"kubernetes.io/projected/516d476a-1a73-4e62-a06f-d9eb90541500-kube-api-access-n4t87\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-4djc2\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.084531 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-utilities\") pod \"certified-operators-s5hlk\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.084612 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2l9p\" (UniqueName: \"kubernetes.io/projected/2286403a-5095-458a-9f73-83b378451237-kube-api-access-j2l9p\") pod \"certified-operators-s5hlk\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.084642 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-catalog-content\") pod \"certified-operators-s5hlk\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.085090 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-catalog-content\") pod \"certified-operators-s5hlk\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.085294 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-utilities\") pod \"certified-operators-s5hlk\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.100105 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2l9p\" (UniqueName: \"kubernetes.io/projected/2286403a-5095-458a-9f73-83b378451237-kube-api-access-j2l9p\") pod \"certified-operators-s5hlk\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.179735 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.276447 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.730747 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2"] Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.750013 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" event={"ID":"516d476a-1a73-4e62-a06f-d9eb90541500","Type":"ContainerStarted","Data":"ac4030624459c8da42406807fa189cd0c5261834db8e421964ef1d12eb149023"} Oct 08 14:38:58 crc kubenswrapper[4789]: I1008 14:38:58.813974 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s5hlk"] Oct 08 14:38:58 crc kubenswrapper[4789]: W1008 14:38:58.816032 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2286403a_5095_458a_9f73_83b378451237.slice/crio-3a10a1b9336d904f1a740f053965093832734013f9ab21504edfb38a7ac1ec76 WatchSource:0}: Error finding container 3a10a1b9336d904f1a740f053965093832734013f9ab21504edfb38a7ac1ec76: Status 404 returned error can't find the container with id 3a10a1b9336d904f1a740f053965093832734013f9ab21504edfb38a7ac1ec76 Oct 08 14:38:59 crc kubenswrapper[4789]: I1008 14:38:59.759258 4789 generic.go:334] "Generic (PLEG): container finished" podID="2286403a-5095-458a-9f73-83b378451237" containerID="bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992" exitCode=0 Oct 08 14:38:59 crc kubenswrapper[4789]: I1008 14:38:59.759366 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5hlk" event={"ID":"2286403a-5095-458a-9f73-83b378451237","Type":"ContainerDied","Data":"bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992"} Oct 08 14:38:59 crc kubenswrapper[4789]: I1008 14:38:59.759621 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5hlk" event={"ID":"2286403a-5095-458a-9f73-83b378451237","Type":"ContainerStarted","Data":"3a10a1b9336d904f1a740f053965093832734013f9ab21504edfb38a7ac1ec76"} Oct 08 14:38:59 crc kubenswrapper[4789]: I1008 14:38:59.761586 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" event={"ID":"516d476a-1a73-4e62-a06f-d9eb90541500","Type":"ContainerStarted","Data":"eb7255b2e0d4c29b98801ec7c43daa4639e59e5f21061f26c4e9583228e190ea"} Oct 08 14:38:59 crc kubenswrapper[4789]: I1008 14:38:59.801453 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" podStartSLOduration=2.249060073 podStartE2EDuration="2.801430098s" podCreationTimestamp="2025-10-08 14:38:57 +0000 UTC" firstStartedPulling="2025-10-08 14:38:58.738204027 +0000 UTC m=+2278.644951519" lastFinishedPulling="2025-10-08 14:38:59.290574052 +0000 UTC m=+2279.197321544" observedRunningTime="2025-10-08 14:38:59.792804454 +0000 UTC m=+2279.699551956" watchObservedRunningTime="2025-10-08 14:38:59.801430098 +0000 UTC m=+2279.708177590" Oct 08 14:39:00 crc kubenswrapper[4789]: I1008 14:39:00.773589 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5hlk" event={"ID":"2286403a-5095-458a-9f73-83b378451237","Type":"ContainerStarted","Data":"decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680"} Oct 08 14:39:01 crc kubenswrapper[4789]: I1008 14:39:01.786387 4789 generic.go:334] "Generic (PLEG): container finished" podID="2286403a-5095-458a-9f73-83b378451237" containerID="decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680" exitCode=0 Oct 08 14:39:01 crc kubenswrapper[4789]: I1008 14:39:01.786449 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5hlk" event={"ID":"2286403a-5095-458a-9f73-83b378451237","Type":"ContainerDied","Data":"decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680"} Oct 08 14:39:02 crc kubenswrapper[4789]: I1008 14:39:02.797297 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5hlk" event={"ID":"2286403a-5095-458a-9f73-83b378451237","Type":"ContainerStarted","Data":"1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e"} Oct 08 14:39:02 crc kubenswrapper[4789]: I1008 14:39:02.822906 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s5hlk" podStartSLOduration=3.241821427 podStartE2EDuration="5.822886675s" podCreationTimestamp="2025-10-08 14:38:57 +0000 UTC" firstStartedPulling="2025-10-08 14:38:59.762127399 +0000 UTC m=+2279.668874891" lastFinishedPulling="2025-10-08 14:39:02.343192647 +0000 UTC m=+2282.249940139" observedRunningTime="2025-10-08 14:39:02.814150218 +0000 UTC m=+2282.720897720" watchObservedRunningTime="2025-10-08 14:39:02.822886675 +0000 UTC m=+2282.729634167" Oct 08 14:39:05 crc kubenswrapper[4789]: I1008 14:39:05.730337 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:39:05 crc kubenswrapper[4789]: E1008 14:39:05.731099 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:39:08 crc kubenswrapper[4789]: I1008 14:39:08.277970 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:39:08 crc kubenswrapper[4789]: I1008 14:39:08.278308 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:39:08 crc kubenswrapper[4789]: I1008 14:39:08.328466 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:39:08 crc kubenswrapper[4789]: I1008 14:39:08.939729 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:39:08 crc kubenswrapper[4789]: I1008 14:39:08.982934 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s5hlk"] Oct 08 14:39:10 crc kubenswrapper[4789]: I1008 14:39:10.907534 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s5hlk" podUID="2286403a-5095-458a-9f73-83b378451237" containerName="registry-server" containerID="cri-o://1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e" gracePeriod=2 Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.386439 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.573085 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-catalog-content\") pod \"2286403a-5095-458a-9f73-83b378451237\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.573356 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-utilities\") pod \"2286403a-5095-458a-9f73-83b378451237\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.573470 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2l9p\" (UniqueName: \"kubernetes.io/projected/2286403a-5095-458a-9f73-83b378451237-kube-api-access-j2l9p\") pod \"2286403a-5095-458a-9f73-83b378451237\" (UID: \"2286403a-5095-458a-9f73-83b378451237\") " Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.574505 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-utilities" (OuterVolumeSpecName: "utilities") pod "2286403a-5095-458a-9f73-83b378451237" (UID: "2286403a-5095-458a-9f73-83b378451237"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.580080 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2286403a-5095-458a-9f73-83b378451237-kube-api-access-j2l9p" (OuterVolumeSpecName: "kube-api-access-j2l9p") pod "2286403a-5095-458a-9f73-83b378451237" (UID: "2286403a-5095-458a-9f73-83b378451237"). InnerVolumeSpecName "kube-api-access-j2l9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.676276 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.676308 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2l9p\" (UniqueName: \"kubernetes.io/projected/2286403a-5095-458a-9f73-83b378451237-kube-api-access-j2l9p\") on node \"crc\" DevicePath \"\"" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.714551 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2286403a-5095-458a-9f73-83b378451237" (UID: "2286403a-5095-458a-9f73-83b378451237"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.777621 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2286403a-5095-458a-9f73-83b378451237-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.921803 4789 generic.go:334] "Generic (PLEG): container finished" podID="2286403a-5095-458a-9f73-83b378451237" containerID="1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e" exitCode=0 Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.921855 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5hlk" event={"ID":"2286403a-5095-458a-9f73-83b378451237","Type":"ContainerDied","Data":"1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e"} Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.921888 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5hlk" event={"ID":"2286403a-5095-458a-9f73-83b378451237","Type":"ContainerDied","Data":"3a10a1b9336d904f1a740f053965093832734013f9ab21504edfb38a7ac1ec76"} Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.921908 4789 scope.go:117] "RemoveContainer" containerID="1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.922069 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5hlk" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.951623 4789 scope.go:117] "RemoveContainer" containerID="decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.968800 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s5hlk"] Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.987778 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s5hlk"] Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.993034 4789 scope.go:117] "RemoveContainer" containerID="bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992" Oct 08 14:39:11 crc kubenswrapper[4789]: I1008 14:39:11.997912 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qwc9b"] Oct 08 14:39:12 crc kubenswrapper[4789]: E1008 14:39:12.010235 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2286403a-5095-458a-9f73-83b378451237" containerName="extract-content" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.010267 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2286403a-5095-458a-9f73-83b378451237" containerName="extract-content" Oct 08 14:39:12 crc kubenswrapper[4789]: E1008 14:39:12.010282 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2286403a-5095-458a-9f73-83b378451237" containerName="registry-server" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.010288 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2286403a-5095-458a-9f73-83b378451237" containerName="registry-server" Oct 08 14:39:12 crc kubenswrapper[4789]: E1008 14:39:12.010307 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2286403a-5095-458a-9f73-83b378451237" containerName="extract-utilities" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.010315 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2286403a-5095-458a-9f73-83b378451237" containerName="extract-utilities" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.010501 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2286403a-5095-458a-9f73-83b378451237" containerName="registry-server" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.014608 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwc9b"] Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.015018 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.027396 4789 scope.go:117] "RemoveContainer" containerID="1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e" Oct 08 14:39:12 crc kubenswrapper[4789]: E1008 14:39:12.030386 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e\": container with ID starting with 1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e not found: ID does not exist" containerID="1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.030533 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e"} err="failed to get container status \"1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e\": rpc error: code = NotFound desc = could not find container \"1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e\": container with ID starting with 1ca9a75d87b1b4337fd22530c53ad9b5c4842cb648a47685e091ff422d3e185e not found: ID does not exist" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.030639 4789 scope.go:117] "RemoveContainer" containerID="decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680" Oct 08 14:39:12 crc kubenswrapper[4789]: E1008 14:39:12.031165 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680\": container with ID starting with decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680 not found: ID does not exist" containerID="decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.031264 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680"} err="failed to get container status \"decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680\": rpc error: code = NotFound desc = could not find container \"decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680\": container with ID starting with decb3a334a47079a53403bb1aedfd3c4dc3e1a45e701a0fda2ca560c77544680 not found: ID does not exist" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.031362 4789 scope.go:117] "RemoveContainer" containerID="bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992" Oct 08 14:39:12 crc kubenswrapper[4789]: E1008 14:39:12.031666 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992\": container with ID starting with bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992 not found: ID does not exist" containerID="bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.031714 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992"} err="failed to get container status \"bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992\": rpc error: code = NotFound desc = could not find container \"bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992\": container with ID starting with bd98239539f9b1f948e1484ecee91bcc61ff244cc08d3ac396aa68503150a992 not found: ID does not exist" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.186346 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-catalog-content\") pod \"redhat-operators-qwc9b\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.186654 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-utilities\") pod \"redhat-operators-qwc9b\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.186764 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h88tz\" (UniqueName: \"kubernetes.io/projected/9896f717-c174-4271-bedc-fc3d65f37c76-kube-api-access-h88tz\") pod \"redhat-operators-qwc9b\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.287920 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h88tz\" (UniqueName: \"kubernetes.io/projected/9896f717-c174-4271-bedc-fc3d65f37c76-kube-api-access-h88tz\") pod \"redhat-operators-qwc9b\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.288025 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-catalog-content\") pod \"redhat-operators-qwc9b\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.288054 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-utilities\") pod \"redhat-operators-qwc9b\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.293822 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-catalog-content\") pod \"redhat-operators-qwc9b\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.300769 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-utilities\") pod \"redhat-operators-qwc9b\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.325826 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h88tz\" (UniqueName: \"kubernetes.io/projected/9896f717-c174-4271-bedc-fc3d65f37c76-kube-api-access-h88tz\") pod \"redhat-operators-qwc9b\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.393793 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.752191 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2286403a-5095-458a-9f73-83b378451237" path="/var/lib/kubelet/pods/2286403a-5095-458a-9f73-83b378451237/volumes" Oct 08 14:39:12 crc kubenswrapper[4789]: W1008 14:39:12.832082 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9896f717_c174_4271_bedc_fc3d65f37c76.slice/crio-c78555c003bafba158e48521c700e49794c4765f00b448d0eea65c074fce346f WatchSource:0}: Error finding container c78555c003bafba158e48521c700e49794c4765f00b448d0eea65c074fce346f: Status 404 returned error can't find the container with id c78555c003bafba158e48521c700e49794c4765f00b448d0eea65c074fce346f Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.843587 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwc9b"] Oct 08 14:39:12 crc kubenswrapper[4789]: I1008 14:39:12.933327 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwc9b" event={"ID":"9896f717-c174-4271-bedc-fc3d65f37c76","Type":"ContainerStarted","Data":"c78555c003bafba158e48521c700e49794c4765f00b448d0eea65c074fce346f"} Oct 08 14:39:13 crc kubenswrapper[4789]: I1008 14:39:13.943906 4789 generic.go:334] "Generic (PLEG): container finished" podID="9896f717-c174-4271-bedc-fc3d65f37c76" containerID="9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f" exitCode=0 Oct 08 14:39:13 crc kubenswrapper[4789]: I1008 14:39:13.944091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwc9b" event={"ID":"9896f717-c174-4271-bedc-fc3d65f37c76","Type":"ContainerDied","Data":"9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f"} Oct 08 14:39:14 crc kubenswrapper[4789]: I1008 14:39:14.964682 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwc9b" event={"ID":"9896f717-c174-4271-bedc-fc3d65f37c76","Type":"ContainerStarted","Data":"f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27"} Oct 08 14:39:15 crc kubenswrapper[4789]: I1008 14:39:15.977077 4789 generic.go:334] "Generic (PLEG): container finished" podID="9896f717-c174-4271-bedc-fc3d65f37c76" containerID="f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27" exitCode=0 Oct 08 14:39:15 crc kubenswrapper[4789]: I1008 14:39:15.977213 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwc9b" event={"ID":"9896f717-c174-4271-bedc-fc3d65f37c76","Type":"ContainerDied","Data":"f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27"} Oct 08 14:39:16 crc kubenswrapper[4789]: I1008 14:39:16.988472 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwc9b" event={"ID":"9896f717-c174-4271-bedc-fc3d65f37c76","Type":"ContainerStarted","Data":"f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9"} Oct 08 14:39:17 crc kubenswrapper[4789]: I1008 14:39:17.010771 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qwc9b" podStartSLOduration=3.25826662 podStartE2EDuration="6.010754202s" podCreationTimestamp="2025-10-08 14:39:11 +0000 UTC" firstStartedPulling="2025-10-08 14:39:13.946387057 +0000 UTC m=+2293.853134569" lastFinishedPulling="2025-10-08 14:39:16.698874659 +0000 UTC m=+2296.605622151" observedRunningTime="2025-10-08 14:39:17.005644753 +0000 UTC m=+2296.912392255" watchObservedRunningTime="2025-10-08 14:39:17.010754202 +0000 UTC m=+2296.917501694" Oct 08 14:39:18 crc kubenswrapper[4789]: I1008 14:39:18.730806 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:39:18 crc kubenswrapper[4789]: E1008 14:39:18.731106 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:39:22 crc kubenswrapper[4789]: I1008 14:39:22.394801 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:22 crc kubenswrapper[4789]: I1008 14:39:22.395418 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:22 crc kubenswrapper[4789]: I1008 14:39:22.450083 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:23 crc kubenswrapper[4789]: I1008 14:39:23.097833 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:23 crc kubenswrapper[4789]: I1008 14:39:23.146196 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwc9b"] Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.083132 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qwc9b" podUID="9896f717-c174-4271-bedc-fc3d65f37c76" containerName="registry-server" containerID="cri-o://f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9" gracePeriod=2 Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.572759 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.643813 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-utilities\") pod \"9896f717-c174-4271-bedc-fc3d65f37c76\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.643929 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h88tz\" (UniqueName: \"kubernetes.io/projected/9896f717-c174-4271-bedc-fc3d65f37c76-kube-api-access-h88tz\") pod \"9896f717-c174-4271-bedc-fc3d65f37c76\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.644054 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-catalog-content\") pod \"9896f717-c174-4271-bedc-fc3d65f37c76\" (UID: \"9896f717-c174-4271-bedc-fc3d65f37c76\") " Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.644935 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-utilities" (OuterVolumeSpecName: "utilities") pod "9896f717-c174-4271-bedc-fc3d65f37c76" (UID: "9896f717-c174-4271-bedc-fc3d65f37c76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.649229 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9896f717-c174-4271-bedc-fc3d65f37c76-kube-api-access-h88tz" (OuterVolumeSpecName: "kube-api-access-h88tz") pod "9896f717-c174-4271-bedc-fc3d65f37c76" (UID: "9896f717-c174-4271-bedc-fc3d65f37c76"). InnerVolumeSpecName "kube-api-access-h88tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.746403 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.746436 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h88tz\" (UniqueName: \"kubernetes.io/projected/9896f717-c174-4271-bedc-fc3d65f37c76-kube-api-access-h88tz\") on node \"crc\" DevicePath \"\"" Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.922610 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9896f717-c174-4271-bedc-fc3d65f37c76" (UID: "9896f717-c174-4271-bedc-fc3d65f37c76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:39:25 crc kubenswrapper[4789]: I1008 14:39:25.950865 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9896f717-c174-4271-bedc-fc3d65f37c76-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.092366 4789 generic.go:334] "Generic (PLEG): container finished" podID="9896f717-c174-4271-bedc-fc3d65f37c76" containerID="f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9" exitCode=0 Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.092462 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwc9b" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.094048 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwc9b" event={"ID":"9896f717-c174-4271-bedc-fc3d65f37c76","Type":"ContainerDied","Data":"f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9"} Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.094144 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwc9b" event={"ID":"9896f717-c174-4271-bedc-fc3d65f37c76","Type":"ContainerDied","Data":"c78555c003bafba158e48521c700e49794c4765f00b448d0eea65c074fce346f"} Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.094185 4789 scope.go:117] "RemoveContainer" containerID="f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.115553 4789 scope.go:117] "RemoveContainer" containerID="f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.137090 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwc9b"] Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.147646 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qwc9b"] Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.150294 4789 scope.go:117] "RemoveContainer" containerID="9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.185279 4789 scope.go:117] "RemoveContainer" containerID="f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9" Oct 08 14:39:26 crc kubenswrapper[4789]: E1008 14:39:26.185705 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9\": container with ID starting with f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9 not found: ID does not exist" containerID="f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.185754 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9"} err="failed to get container status \"f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9\": rpc error: code = NotFound desc = could not find container \"f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9\": container with ID starting with f6c639f0a787d579584379d020d69dd682dc24b7a88ccba26a8c2931a1a83aa9 not found: ID does not exist" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.185777 4789 scope.go:117] "RemoveContainer" containerID="f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27" Oct 08 14:39:26 crc kubenswrapper[4789]: E1008 14:39:26.186115 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27\": container with ID starting with f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27 not found: ID does not exist" containerID="f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.186261 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27"} err="failed to get container status \"f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27\": rpc error: code = NotFound desc = could not find container \"f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27\": container with ID starting with f3bfcaa0fcbb2814af6077a387839e24ec4c601d05c4a49014f3e13664ee7b27 not found: ID does not exist" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.186380 4789 scope.go:117] "RemoveContainer" containerID="9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f" Oct 08 14:39:26 crc kubenswrapper[4789]: E1008 14:39:26.186735 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f\": container with ID starting with 9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f not found: ID does not exist" containerID="9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.186758 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f"} err="failed to get container status \"9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f\": rpc error: code = NotFound desc = could not find container \"9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f\": container with ID starting with 9c88b459ee7003a4e9696d152308439f2dcef7f961cd09dddba8c24da3d4865f not found: ID does not exist" Oct 08 14:39:26 crc kubenswrapper[4789]: I1008 14:39:26.742946 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9896f717-c174-4271-bedc-fc3d65f37c76" path="/var/lib/kubelet/pods/9896f717-c174-4271-bedc-fc3d65f37c76/volumes" Oct 08 14:39:32 crc kubenswrapper[4789]: I1008 14:39:32.730966 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:39:32 crc kubenswrapper[4789]: E1008 14:39:32.732761 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:39:47 crc kubenswrapper[4789]: I1008 14:39:47.730698 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:39:47 crc kubenswrapper[4789]: E1008 14:39:47.732509 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:40:01 crc kubenswrapper[4789]: I1008 14:40:01.729981 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:40:01 crc kubenswrapper[4789]: E1008 14:40:01.730816 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:40:06 crc kubenswrapper[4789]: I1008 14:40:06.446376 4789 generic.go:334] "Generic (PLEG): container finished" podID="516d476a-1a73-4e62-a06f-d9eb90541500" containerID="eb7255b2e0d4c29b98801ec7c43daa4639e59e5f21061f26c4e9583228e190ea" exitCode=0 Oct 08 14:40:06 crc kubenswrapper[4789]: I1008 14:40:06.446415 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" event={"ID":"516d476a-1a73-4e62-a06f-d9eb90541500","Type":"ContainerDied","Data":"eb7255b2e0d4c29b98801ec7c43daa4639e59e5f21061f26c4e9583228e190ea"} Oct 08 14:40:07 crc kubenswrapper[4789]: I1008 14:40:07.897100 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:40:07 crc kubenswrapper[4789]: I1008 14:40:07.993513 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4t87\" (UniqueName: \"kubernetes.io/projected/516d476a-1a73-4e62-a06f-d9eb90541500-kube-api-access-n4t87\") pod \"516d476a-1a73-4e62-a06f-d9eb90541500\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " Oct 08 14:40:07 crc kubenswrapper[4789]: I1008 14:40:07.994025 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/516d476a-1a73-4e62-a06f-d9eb90541500-ovncontroller-config-0\") pod \"516d476a-1a73-4e62-a06f-d9eb90541500\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " Oct 08 14:40:07 crc kubenswrapper[4789]: I1008 14:40:07.994141 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ovn-combined-ca-bundle\") pod \"516d476a-1a73-4e62-a06f-d9eb90541500\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " Oct 08 14:40:07 crc kubenswrapper[4789]: I1008 14:40:07.994183 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ssh-key\") pod \"516d476a-1a73-4e62-a06f-d9eb90541500\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " Oct 08 14:40:07 crc kubenswrapper[4789]: I1008 14:40:07.994222 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-inventory\") pod \"516d476a-1a73-4e62-a06f-d9eb90541500\" (UID: \"516d476a-1a73-4e62-a06f-d9eb90541500\") " Oct 08 14:40:07 crc kubenswrapper[4789]: I1008 14:40:07.999689 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "516d476a-1a73-4e62-a06f-d9eb90541500" (UID: "516d476a-1a73-4e62-a06f-d9eb90541500"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.004190 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/516d476a-1a73-4e62-a06f-d9eb90541500-kube-api-access-n4t87" (OuterVolumeSpecName: "kube-api-access-n4t87") pod "516d476a-1a73-4e62-a06f-d9eb90541500" (UID: "516d476a-1a73-4e62-a06f-d9eb90541500"). InnerVolumeSpecName "kube-api-access-n4t87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.020584 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-inventory" (OuterVolumeSpecName: "inventory") pod "516d476a-1a73-4e62-a06f-d9eb90541500" (UID: "516d476a-1a73-4e62-a06f-d9eb90541500"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.021958 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "516d476a-1a73-4e62-a06f-d9eb90541500" (UID: "516d476a-1a73-4e62-a06f-d9eb90541500"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.035744 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/516d476a-1a73-4e62-a06f-d9eb90541500-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "516d476a-1a73-4e62-a06f-d9eb90541500" (UID: "516d476a-1a73-4e62-a06f-d9eb90541500"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.096286 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4t87\" (UniqueName: \"kubernetes.io/projected/516d476a-1a73-4e62-a06f-d9eb90541500-kube-api-access-n4t87\") on node \"crc\" DevicePath \"\"" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.096325 4789 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/516d476a-1a73-4e62-a06f-d9eb90541500-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.096334 4789 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.096343 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.096351 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/516d476a-1a73-4e62-a06f-d9eb90541500-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.467251 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" event={"ID":"516d476a-1a73-4e62-a06f-d9eb90541500","Type":"ContainerDied","Data":"ac4030624459c8da42406807fa189cd0c5261834db8e421964ef1d12eb149023"} Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.467291 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac4030624459c8da42406807fa189cd0c5261834db8e421964ef1d12eb149023" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.467357 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-4djc2" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.623964 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt"] Oct 08 14:40:08 crc kubenswrapper[4789]: E1008 14:40:08.624459 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516d476a-1a73-4e62-a06f-d9eb90541500" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.624483 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="516d476a-1a73-4e62-a06f-d9eb90541500" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 14:40:08 crc kubenswrapper[4789]: E1008 14:40:08.624536 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9896f717-c174-4271-bedc-fc3d65f37c76" containerName="extract-utilities" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.624546 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9896f717-c174-4271-bedc-fc3d65f37c76" containerName="extract-utilities" Oct 08 14:40:08 crc kubenswrapper[4789]: E1008 14:40:08.624558 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9896f717-c174-4271-bedc-fc3d65f37c76" containerName="extract-content" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.624565 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9896f717-c174-4271-bedc-fc3d65f37c76" containerName="extract-content" Oct 08 14:40:08 crc kubenswrapper[4789]: E1008 14:40:08.624580 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9896f717-c174-4271-bedc-fc3d65f37c76" containerName="registry-server" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.624589 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="9896f717-c174-4271-bedc-fc3d65f37c76" containerName="registry-server" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.624806 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="516d476a-1a73-4e62-a06f-d9eb90541500" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.624822 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="9896f717-c174-4271-bedc-fc3d65f37c76" containerName="registry-server" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.625711 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.631551 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.631649 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.631703 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.631660 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.631846 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.632050 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.634469 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt"] Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.808265 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.808655 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.808897 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2hbn\" (UniqueName: \"kubernetes.io/projected/7f2ef2f7-aa9f-4767-a032-e461dde85655-kube-api-access-w2hbn\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.809077 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.809236 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.809349 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.911611 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.912779 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.913192 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2hbn\" (UniqueName: \"kubernetes.io/projected/7f2ef2f7-aa9f-4767-a032-e461dde85655-kube-api-access-w2hbn\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.913308 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.913451 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.913548 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.916963 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.917003 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.920369 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.926591 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.934135 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.934966 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2hbn\" (UniqueName: \"kubernetes.io/projected/7f2ef2f7-aa9f-4767-a032-e461dde85655-kube-api-access-w2hbn\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:08 crc kubenswrapper[4789]: I1008 14:40:08.944485 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:40:09 crc kubenswrapper[4789]: W1008 14:40:09.579039 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f2ef2f7_aa9f_4767_a032_e461dde85655.slice/crio-c881179110d487069ce805e5c076c7bf321f01c395d13f70ba95276a56238099 WatchSource:0}: Error finding container c881179110d487069ce805e5c076c7bf321f01c395d13f70ba95276a56238099: Status 404 returned error can't find the container with id c881179110d487069ce805e5c076c7bf321f01c395d13f70ba95276a56238099 Oct 08 14:40:09 crc kubenswrapper[4789]: I1008 14:40:09.583806 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt"] Oct 08 14:40:10 crc kubenswrapper[4789]: I1008 14:40:10.493051 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" event={"ID":"7f2ef2f7-aa9f-4767-a032-e461dde85655","Type":"ContainerStarted","Data":"ec23e68e7c296e9ae0312f70cd7733161fbea418390bb6157eb4f9a105cbde14"} Oct 08 14:40:10 crc kubenswrapper[4789]: I1008 14:40:10.493594 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" event={"ID":"7f2ef2f7-aa9f-4767-a032-e461dde85655","Type":"ContainerStarted","Data":"c881179110d487069ce805e5c076c7bf321f01c395d13f70ba95276a56238099"} Oct 08 14:40:10 crc kubenswrapper[4789]: I1008 14:40:10.510730 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" podStartSLOduration=2.108319997 podStartE2EDuration="2.510712073s" podCreationTimestamp="2025-10-08 14:40:08 +0000 UTC" firstStartedPulling="2025-10-08 14:40:09.582233947 +0000 UTC m=+2349.488981439" lastFinishedPulling="2025-10-08 14:40:09.984626013 +0000 UTC m=+2349.891373515" observedRunningTime="2025-10-08 14:40:10.509442249 +0000 UTC m=+2350.416189741" watchObservedRunningTime="2025-10-08 14:40:10.510712073 +0000 UTC m=+2350.417459565" Oct 08 14:40:12 crc kubenswrapper[4789]: I1008 14:40:12.731673 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:40:12 crc kubenswrapper[4789]: E1008 14:40:12.732629 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:40:26 crc kubenswrapper[4789]: I1008 14:40:26.730278 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:40:26 crc kubenswrapper[4789]: E1008 14:40:26.731071 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:40:37 crc kubenswrapper[4789]: I1008 14:40:37.730411 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:40:37 crc kubenswrapper[4789]: E1008 14:40:37.732007 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:40:50 crc kubenswrapper[4789]: I1008 14:40:50.738399 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:40:50 crc kubenswrapper[4789]: E1008 14:40:50.739169 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:41:00 crc kubenswrapper[4789]: I1008 14:41:00.969694 4789 generic.go:334] "Generic (PLEG): container finished" podID="7f2ef2f7-aa9f-4767-a032-e461dde85655" containerID="ec23e68e7c296e9ae0312f70cd7733161fbea418390bb6157eb4f9a105cbde14" exitCode=0 Oct 08 14:41:00 crc kubenswrapper[4789]: I1008 14:41:00.969870 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" event={"ID":"7f2ef2f7-aa9f-4767-a032-e461dde85655","Type":"ContainerDied","Data":"ec23e68e7c296e9ae0312f70cd7733161fbea418390bb6157eb4f9a105cbde14"} Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.418541 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.566735 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2hbn\" (UniqueName: \"kubernetes.io/projected/7f2ef2f7-aa9f-4767-a032-e461dde85655-kube-api-access-w2hbn\") pod \"7f2ef2f7-aa9f-4767-a032-e461dde85655\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.566828 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-nova-metadata-neutron-config-0\") pod \"7f2ef2f7-aa9f-4767-a032-e461dde85655\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.566864 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-ovn-metadata-agent-neutron-config-0\") pod \"7f2ef2f7-aa9f-4767-a032-e461dde85655\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.567027 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-inventory\") pod \"7f2ef2f7-aa9f-4767-a032-e461dde85655\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.567077 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-ssh-key\") pod \"7f2ef2f7-aa9f-4767-a032-e461dde85655\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.567156 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-metadata-combined-ca-bundle\") pod \"7f2ef2f7-aa9f-4767-a032-e461dde85655\" (UID: \"7f2ef2f7-aa9f-4767-a032-e461dde85655\") " Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.573340 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f2ef2f7-aa9f-4767-a032-e461dde85655-kube-api-access-w2hbn" (OuterVolumeSpecName: "kube-api-access-w2hbn") pod "7f2ef2f7-aa9f-4767-a032-e461dde85655" (UID: "7f2ef2f7-aa9f-4767-a032-e461dde85655"). InnerVolumeSpecName "kube-api-access-w2hbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.574406 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "7f2ef2f7-aa9f-4767-a032-e461dde85655" (UID: "7f2ef2f7-aa9f-4767-a032-e461dde85655"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.645191 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "7f2ef2f7-aa9f-4767-a032-e461dde85655" (UID: "7f2ef2f7-aa9f-4767-a032-e461dde85655"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.658120 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-inventory" (OuterVolumeSpecName: "inventory") pod "7f2ef2f7-aa9f-4767-a032-e461dde85655" (UID: "7f2ef2f7-aa9f-4767-a032-e461dde85655"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.669425 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2hbn\" (UniqueName: \"kubernetes.io/projected/7f2ef2f7-aa9f-4767-a032-e461dde85655-kube-api-access-w2hbn\") on node \"crc\" DevicePath \"\"" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.669461 4789 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.669473 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.669482 4789 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.681775 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "7f2ef2f7-aa9f-4767-a032-e461dde85655" (UID: "7f2ef2f7-aa9f-4767-a032-e461dde85655"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.689156 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7f2ef2f7-aa9f-4767-a032-e461dde85655" (UID: "7f2ef2f7-aa9f-4767-a032-e461dde85655"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.732934 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:41:02 crc kubenswrapper[4789]: E1008 14:41:02.733286 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.770799 4789 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:41:02 crc kubenswrapper[4789]: I1008 14:41:02.770839 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f2ef2f7-aa9f-4767-a032-e461dde85655-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:41:02 crc kubenswrapper[4789]: E1008 14:41:02.982607 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f2ef2f7_aa9f_4767_a032_e461dde85655.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f2ef2f7_aa9f_4767_a032_e461dde85655.slice/crio-c881179110d487069ce805e5c076c7bf321f01c395d13f70ba95276a56238099\": RecentStats: unable to find data in memory cache]" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.000272 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" event={"ID":"7f2ef2f7-aa9f-4767-a032-e461dde85655","Type":"ContainerDied","Data":"c881179110d487069ce805e5c076c7bf321f01c395d13f70ba95276a56238099"} Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.000574 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c881179110d487069ce805e5c076c7bf321f01c395d13f70ba95276a56238099" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.000697 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.078090 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg"] Oct 08 14:41:03 crc kubenswrapper[4789]: E1008 14:41:03.078677 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f2ef2f7-aa9f-4767-a032-e461dde85655" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.078751 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f2ef2f7-aa9f-4767-a032-e461dde85655" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.079081 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f2ef2f7-aa9f-4767-a032-e461dde85655" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.079939 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.087166 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg"] Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.088080 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.088322 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.088519 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.088734 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.089108 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.177039 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.177512 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-489c9\" (UniqueName: \"kubernetes.io/projected/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-kube-api-access-489c9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.177592 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.177690 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.177839 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.295630 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-489c9\" (UniqueName: \"kubernetes.io/projected/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-kube-api-access-489c9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.295675 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.295702 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.295725 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.295803 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.300778 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.300872 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.300865 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.311217 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.319847 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-489c9\" (UniqueName: \"kubernetes.io/projected/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-kube-api-access-489c9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.395911 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:41:03 crc kubenswrapper[4789]: I1008 14:41:03.928388 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg"] Oct 08 14:41:04 crc kubenswrapper[4789]: I1008 14:41:04.011343 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" event={"ID":"2066f52b-bb99-4a61-98e4-63f53b8ec7c7","Type":"ContainerStarted","Data":"156f372cb8656337666844a6a8c8379c4dba484672b7d930eefc7ed92a742839"} Oct 08 14:41:05 crc kubenswrapper[4789]: I1008 14:41:05.021065 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" event={"ID":"2066f52b-bb99-4a61-98e4-63f53b8ec7c7","Type":"ContainerStarted","Data":"f534a4fbf1eaecc252e5c1f358812952f6aa2f427b235f4802d9141ed04b08cc"} Oct 08 14:41:05 crc kubenswrapper[4789]: I1008 14:41:05.038735 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" podStartSLOduration=1.4919452899999999 podStartE2EDuration="2.038714429s" podCreationTimestamp="2025-10-08 14:41:03 +0000 UTC" firstStartedPulling="2025-10-08 14:41:03.930370083 +0000 UTC m=+2403.837117585" lastFinishedPulling="2025-10-08 14:41:04.477139232 +0000 UTC m=+2404.383886724" observedRunningTime="2025-10-08 14:41:05.035015467 +0000 UTC m=+2404.941762969" watchObservedRunningTime="2025-10-08 14:41:05.038714429 +0000 UTC m=+2404.945461941" Oct 08 14:41:13 crc kubenswrapper[4789]: I1008 14:41:13.731835 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:41:13 crc kubenswrapper[4789]: E1008 14:41:13.733299 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:41:25 crc kubenswrapper[4789]: I1008 14:41:25.730093 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:41:25 crc kubenswrapper[4789]: E1008 14:41:25.730965 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:41:39 crc kubenswrapper[4789]: I1008 14:41:39.729762 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:41:39 crc kubenswrapper[4789]: E1008 14:41:39.730422 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:41:53 crc kubenswrapper[4789]: I1008 14:41:53.730176 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:41:53 crc kubenswrapper[4789]: E1008 14:41:53.731776 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:42:07 crc kubenswrapper[4789]: I1008 14:42:07.730493 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:42:07 crc kubenswrapper[4789]: E1008 14:42:07.731271 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:42:19 crc kubenswrapper[4789]: I1008 14:42:19.730483 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:42:19 crc kubenswrapper[4789]: E1008 14:42:19.731657 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:42:34 crc kubenswrapper[4789]: I1008 14:42:34.729743 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:42:34 crc kubenswrapper[4789]: E1008 14:42:34.730472 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:42:48 crc kubenswrapper[4789]: I1008 14:42:48.731961 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:42:48 crc kubenswrapper[4789]: E1008 14:42:48.732804 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:43:00 crc kubenswrapper[4789]: I1008 14:43:00.743746 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:43:00 crc kubenswrapper[4789]: E1008 14:43:00.744939 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:43:11 crc kubenswrapper[4789]: I1008 14:43:11.730650 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:43:11 crc kubenswrapper[4789]: E1008 14:43:11.731779 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:43:25 crc kubenswrapper[4789]: I1008 14:43:25.730242 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:43:25 crc kubenswrapper[4789]: E1008 14:43:25.731081 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:43:40 crc kubenswrapper[4789]: I1008 14:43:40.736446 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:43:41 crc kubenswrapper[4789]: I1008 14:43:41.549881 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"5f903459af49810b12480e53c24b2593b65c7aac101fb02e8a672f71a5607dbb"} Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.156481 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb"] Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.168793 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.193758 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.195728 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.201340 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb"] Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.229416 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rwdr\" (UniqueName: \"kubernetes.io/projected/8edae8fb-3e38-477a-aa7f-5f6746833c47-kube-api-access-7rwdr\") pod \"collect-profiles-29332245-fk2nb\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.229593 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edae8fb-3e38-477a-aa7f-5f6746833c47-secret-volume\") pod \"collect-profiles-29332245-fk2nb\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.229711 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edae8fb-3e38-477a-aa7f-5f6746833c47-config-volume\") pod \"collect-profiles-29332245-fk2nb\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.333107 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edae8fb-3e38-477a-aa7f-5f6746833c47-config-volume\") pod \"collect-profiles-29332245-fk2nb\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.333228 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rwdr\" (UniqueName: \"kubernetes.io/projected/8edae8fb-3e38-477a-aa7f-5f6746833c47-kube-api-access-7rwdr\") pod \"collect-profiles-29332245-fk2nb\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.333305 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edae8fb-3e38-477a-aa7f-5f6746833c47-secret-volume\") pod \"collect-profiles-29332245-fk2nb\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.334909 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edae8fb-3e38-477a-aa7f-5f6746833c47-config-volume\") pod \"collect-profiles-29332245-fk2nb\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.356499 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edae8fb-3e38-477a-aa7f-5f6746833c47-secret-volume\") pod \"collect-profiles-29332245-fk2nb\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.366797 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rwdr\" (UniqueName: \"kubernetes.io/projected/8edae8fb-3e38-477a-aa7f-5f6746833c47-kube-api-access-7rwdr\") pod \"collect-profiles-29332245-fk2nb\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.525585 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:00 crc kubenswrapper[4789]: I1008 14:45:00.977386 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb"] Oct 08 14:45:01 crc kubenswrapper[4789]: I1008 14:45:01.385073 4789 generic.go:334] "Generic (PLEG): container finished" podID="8edae8fb-3e38-477a-aa7f-5f6746833c47" containerID="fb9bf00f8831ca940e98c3aadfe9d5641c3f9184fe4e80c4d7254b1e08190e2a" exitCode=0 Oct 08 14:45:01 crc kubenswrapper[4789]: I1008 14:45:01.385125 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" event={"ID":"8edae8fb-3e38-477a-aa7f-5f6746833c47","Type":"ContainerDied","Data":"fb9bf00f8831ca940e98c3aadfe9d5641c3f9184fe4e80c4d7254b1e08190e2a"} Oct 08 14:45:01 crc kubenswrapper[4789]: I1008 14:45:01.385195 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" event={"ID":"8edae8fb-3e38-477a-aa7f-5f6746833c47","Type":"ContainerStarted","Data":"d6e143fa82f7df10e311bbc405706ca8bce49021eae16fe97e5556bf0e56c82e"} Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.739892 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.877651 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edae8fb-3e38-477a-aa7f-5f6746833c47-secret-volume\") pod \"8edae8fb-3e38-477a-aa7f-5f6746833c47\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.877927 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edae8fb-3e38-477a-aa7f-5f6746833c47-config-volume\") pod \"8edae8fb-3e38-477a-aa7f-5f6746833c47\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.877964 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rwdr\" (UniqueName: \"kubernetes.io/projected/8edae8fb-3e38-477a-aa7f-5f6746833c47-kube-api-access-7rwdr\") pod \"8edae8fb-3e38-477a-aa7f-5f6746833c47\" (UID: \"8edae8fb-3e38-477a-aa7f-5f6746833c47\") " Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.879814 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8edae8fb-3e38-477a-aa7f-5f6746833c47-config-volume" (OuterVolumeSpecName: "config-volume") pod "8edae8fb-3e38-477a-aa7f-5f6746833c47" (UID: "8edae8fb-3e38-477a-aa7f-5f6746833c47"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.882844 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8edae8fb-3e38-477a-aa7f-5f6746833c47-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8edae8fb-3e38-477a-aa7f-5f6746833c47" (UID: "8edae8fb-3e38-477a-aa7f-5f6746833c47"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.884222 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8edae8fb-3e38-477a-aa7f-5f6746833c47-kube-api-access-7rwdr" (OuterVolumeSpecName: "kube-api-access-7rwdr") pod "8edae8fb-3e38-477a-aa7f-5f6746833c47" (UID: "8edae8fb-3e38-477a-aa7f-5f6746833c47"). InnerVolumeSpecName "kube-api-access-7rwdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.979709 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8edae8fb-3e38-477a-aa7f-5f6746833c47-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.979744 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8edae8fb-3e38-477a-aa7f-5f6746833c47-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:02 crc kubenswrapper[4789]: I1008 14:45:02.979754 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rwdr\" (UniqueName: \"kubernetes.io/projected/8edae8fb-3e38-477a-aa7f-5f6746833c47-kube-api-access-7rwdr\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:03 crc kubenswrapper[4789]: I1008 14:45:03.406769 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" event={"ID":"8edae8fb-3e38-477a-aa7f-5f6746833c47","Type":"ContainerDied","Data":"d6e143fa82f7df10e311bbc405706ca8bce49021eae16fe97e5556bf0e56c82e"} Oct 08 14:45:03 crc kubenswrapper[4789]: I1008 14:45:03.406808 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6e143fa82f7df10e311bbc405706ca8bce49021eae16fe97e5556bf0e56c82e" Oct 08 14:45:03 crc kubenswrapper[4789]: I1008 14:45:03.406852 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb" Oct 08 14:45:03 crc kubenswrapper[4789]: I1008 14:45:03.836953 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp"] Oct 08 14:45:03 crc kubenswrapper[4789]: I1008 14:45:03.843946 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332200-8zphp"] Oct 08 14:45:04 crc kubenswrapper[4789]: I1008 14:45:04.744975 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2389a066-e146-4ea8-bb92-d596311b564a" path="/var/lib/kubelet/pods/2389a066-e146-4ea8-bb92-d596311b564a/volumes" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.530312 4789 scope.go:117] "RemoveContainer" containerID="a111de65cc85e6d65135e4262b26fdc5335bddda34a02b6473abb9cc6425bb89" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.723133 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2xgjj"] Oct 08 14:45:23 crc kubenswrapper[4789]: E1008 14:45:23.723810 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8edae8fb-3e38-477a-aa7f-5f6746833c47" containerName="collect-profiles" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.723823 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8edae8fb-3e38-477a-aa7f-5f6746833c47" containerName="collect-profiles" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.724088 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8edae8fb-3e38-477a-aa7f-5f6746833c47" containerName="collect-profiles" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.725478 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.732636 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2xgjj"] Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.861870 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-catalog-content\") pod \"community-operators-2xgjj\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.861911 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d4jl\" (UniqueName: \"kubernetes.io/projected/e98951cd-7535-4396-b6db-b1adfb45ce51-kube-api-access-8d4jl\") pod \"community-operators-2xgjj\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.862205 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-utilities\") pod \"community-operators-2xgjj\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.963819 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-utilities\") pod \"community-operators-2xgjj\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.963959 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-catalog-content\") pod \"community-operators-2xgjj\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.964004 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d4jl\" (UniqueName: \"kubernetes.io/projected/e98951cd-7535-4396-b6db-b1adfb45ce51-kube-api-access-8d4jl\") pod \"community-operators-2xgjj\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.964934 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-utilities\") pod \"community-operators-2xgjj\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.965244 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-catalog-content\") pod \"community-operators-2xgjj\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:23 crc kubenswrapper[4789]: I1008 14:45:23.985624 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d4jl\" (UniqueName: \"kubernetes.io/projected/e98951cd-7535-4396-b6db-b1adfb45ce51-kube-api-access-8d4jl\") pod \"community-operators-2xgjj\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:24 crc kubenswrapper[4789]: I1008 14:45:24.053257 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:24 crc kubenswrapper[4789]: I1008 14:45:24.604609 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2xgjj"] Oct 08 14:45:25 crc kubenswrapper[4789]: I1008 14:45:25.593626 4789 generic.go:334] "Generic (PLEG): container finished" podID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerID="df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63" exitCode=0 Oct 08 14:45:25 crc kubenswrapper[4789]: I1008 14:45:25.593734 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xgjj" event={"ID":"e98951cd-7535-4396-b6db-b1adfb45ce51","Type":"ContainerDied","Data":"df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63"} Oct 08 14:45:25 crc kubenswrapper[4789]: I1008 14:45:25.594283 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xgjj" event={"ID":"e98951cd-7535-4396-b6db-b1adfb45ce51","Type":"ContainerStarted","Data":"86dbf8a834f5dbf916ffce184abd47bb3d448398b651800edbf8f448dfab3f4d"} Oct 08 14:45:25 crc kubenswrapper[4789]: I1008 14:45:25.596954 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 14:45:26 crc kubenswrapper[4789]: I1008 14:45:26.604625 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xgjj" event={"ID":"e98951cd-7535-4396-b6db-b1adfb45ce51","Type":"ContainerStarted","Data":"b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f"} Oct 08 14:45:27 crc kubenswrapper[4789]: I1008 14:45:27.614349 4789 generic.go:334] "Generic (PLEG): container finished" podID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerID="b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f" exitCode=0 Oct 08 14:45:27 crc kubenswrapper[4789]: I1008 14:45:27.614397 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xgjj" event={"ID":"e98951cd-7535-4396-b6db-b1adfb45ce51","Type":"ContainerDied","Data":"b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f"} Oct 08 14:45:28 crc kubenswrapper[4789]: I1008 14:45:28.626881 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xgjj" event={"ID":"e98951cd-7535-4396-b6db-b1adfb45ce51","Type":"ContainerStarted","Data":"e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01"} Oct 08 14:45:28 crc kubenswrapper[4789]: I1008 14:45:28.647844 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2xgjj" podStartSLOduration=2.871488571 podStartE2EDuration="5.647830089s" podCreationTimestamp="2025-10-08 14:45:23 +0000 UTC" firstStartedPulling="2025-10-08 14:45:25.596711377 +0000 UTC m=+2665.503458869" lastFinishedPulling="2025-10-08 14:45:28.373052895 +0000 UTC m=+2668.279800387" observedRunningTime="2025-10-08 14:45:28.645007252 +0000 UTC m=+2668.551754744" watchObservedRunningTime="2025-10-08 14:45:28.647830089 +0000 UTC m=+2668.554577581" Oct 08 14:45:34 crc kubenswrapper[4789]: I1008 14:45:34.054049 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:34 crc kubenswrapper[4789]: I1008 14:45:34.054605 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:34 crc kubenswrapper[4789]: I1008 14:45:34.100777 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:34 crc kubenswrapper[4789]: I1008 14:45:34.720025 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:34 crc kubenswrapper[4789]: I1008 14:45:34.778398 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2xgjj"] Oct 08 14:45:36 crc kubenswrapper[4789]: I1008 14:45:36.689858 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2xgjj" podUID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerName="registry-server" containerID="cri-o://e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01" gracePeriod=2 Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.135733 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.226615 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-utilities\") pod \"e98951cd-7535-4396-b6db-b1adfb45ce51\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.226871 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d4jl\" (UniqueName: \"kubernetes.io/projected/e98951cd-7535-4396-b6db-b1adfb45ce51-kube-api-access-8d4jl\") pod \"e98951cd-7535-4396-b6db-b1adfb45ce51\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.226923 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-catalog-content\") pod \"e98951cd-7535-4396-b6db-b1adfb45ce51\" (UID: \"e98951cd-7535-4396-b6db-b1adfb45ce51\") " Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.234675 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-utilities" (OuterVolumeSpecName: "utilities") pod "e98951cd-7535-4396-b6db-b1adfb45ce51" (UID: "e98951cd-7535-4396-b6db-b1adfb45ce51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.256263 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e98951cd-7535-4396-b6db-b1adfb45ce51-kube-api-access-8d4jl" (OuterVolumeSpecName: "kube-api-access-8d4jl") pod "e98951cd-7535-4396-b6db-b1adfb45ce51" (UID: "e98951cd-7535-4396-b6db-b1adfb45ce51"). InnerVolumeSpecName "kube-api-access-8d4jl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.329294 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d4jl\" (UniqueName: \"kubernetes.io/projected/e98951cd-7535-4396-b6db-b1adfb45ce51-kube-api-access-8d4jl\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.329586 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.700083 4789 generic.go:334] "Generic (PLEG): container finished" podID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerID="e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01" exitCode=0 Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.700125 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xgjj" event={"ID":"e98951cd-7535-4396-b6db-b1adfb45ce51","Type":"ContainerDied","Data":"e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01"} Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.701089 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2xgjj" event={"ID":"e98951cd-7535-4396-b6db-b1adfb45ce51","Type":"ContainerDied","Data":"86dbf8a834f5dbf916ffce184abd47bb3d448398b651800edbf8f448dfab3f4d"} Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.700139 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2xgjj" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.701138 4789 scope.go:117] "RemoveContainer" containerID="e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.719626 4789 scope.go:117] "RemoveContainer" containerID="b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.742095 4789 scope.go:117] "RemoveContainer" containerID="df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.780177 4789 scope.go:117] "RemoveContainer" containerID="e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01" Oct 08 14:45:37 crc kubenswrapper[4789]: E1008 14:45:37.780622 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01\": container with ID starting with e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01 not found: ID does not exist" containerID="e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.780667 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01"} err="failed to get container status \"e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01\": rpc error: code = NotFound desc = could not find container \"e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01\": container with ID starting with e332acbe81c3fcbef57e0c05d2b32f38f5ab18a37ceabf7d85d86e9fb64baf01 not found: ID does not exist" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.780693 4789 scope.go:117] "RemoveContainer" containerID="b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f" Oct 08 14:45:37 crc kubenswrapper[4789]: E1008 14:45:37.781019 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f\": container with ID starting with b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f not found: ID does not exist" containerID="b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.781050 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f"} err="failed to get container status \"b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f\": rpc error: code = NotFound desc = could not find container \"b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f\": container with ID starting with b52869e3e4a974536f07f07cdf6b23c78717d0ccc228d4e2d5d5d9e4720c7c6f not found: ID does not exist" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.781066 4789 scope.go:117] "RemoveContainer" containerID="df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63" Oct 08 14:45:37 crc kubenswrapper[4789]: E1008 14:45:37.781511 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63\": container with ID starting with df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63 not found: ID does not exist" containerID="df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.781540 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63"} err="failed to get container status \"df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63\": rpc error: code = NotFound desc = could not find container \"df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63\": container with ID starting with df56d3404fac762d710397bf87804c93cdfbc8dc7a904dd70e967fa9e9938d63 not found: ID does not exist" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.915473 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e98951cd-7535-4396-b6db-b1adfb45ce51" (UID: "e98951cd-7535-4396-b6db-b1adfb45ce51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:45:37 crc kubenswrapper[4789]: I1008 14:45:37.940756 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98951cd-7535-4396-b6db-b1adfb45ce51-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:38 crc kubenswrapper[4789]: I1008 14:45:38.040913 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2xgjj"] Oct 08 14:45:38 crc kubenswrapper[4789]: I1008 14:45:38.050509 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2xgjj"] Oct 08 14:45:38 crc kubenswrapper[4789]: I1008 14:45:38.745977 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e98951cd-7535-4396-b6db-b1adfb45ce51" path="/var/lib/kubelet/pods/e98951cd-7535-4396-b6db-b1adfb45ce51/volumes" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.468290 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kz4nf"] Oct 08 14:45:51 crc kubenswrapper[4789]: E1008 14:45:51.469150 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerName="extract-content" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.469164 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerName="extract-content" Oct 08 14:45:51 crc kubenswrapper[4789]: E1008 14:45:51.469173 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerName="extract-utilities" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.469179 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerName="extract-utilities" Oct 08 14:45:51 crc kubenswrapper[4789]: E1008 14:45:51.469195 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerName="registry-server" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.469202 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerName="registry-server" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.469389 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98951cd-7535-4396-b6db-b1adfb45ce51" containerName="registry-server" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.470792 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.488839 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kz4nf"] Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.506732 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-utilities\") pod \"redhat-marketplace-kz4nf\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.506921 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5kf5\" (UniqueName: \"kubernetes.io/projected/bc8b8740-c44b-4eb9-aa37-4071f377d07e-kube-api-access-d5kf5\") pod \"redhat-marketplace-kz4nf\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.506962 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-catalog-content\") pod \"redhat-marketplace-kz4nf\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.609226 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-utilities\") pod \"redhat-marketplace-kz4nf\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.609416 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5kf5\" (UniqueName: \"kubernetes.io/projected/bc8b8740-c44b-4eb9-aa37-4071f377d07e-kube-api-access-d5kf5\") pod \"redhat-marketplace-kz4nf\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.609458 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-catalog-content\") pod \"redhat-marketplace-kz4nf\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.610136 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-catalog-content\") pod \"redhat-marketplace-kz4nf\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.610456 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-utilities\") pod \"redhat-marketplace-kz4nf\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.635081 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5kf5\" (UniqueName: \"kubernetes.io/projected/bc8b8740-c44b-4eb9-aa37-4071f377d07e-kube-api-access-d5kf5\") pod \"redhat-marketplace-kz4nf\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:51 crc kubenswrapper[4789]: I1008 14:45:51.798697 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:45:52 crc kubenswrapper[4789]: I1008 14:45:52.276841 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kz4nf"] Oct 08 14:45:52 crc kubenswrapper[4789]: I1008 14:45:52.847382 4789 generic.go:334] "Generic (PLEG): container finished" podID="2066f52b-bb99-4a61-98e4-63f53b8ec7c7" containerID="f534a4fbf1eaecc252e5c1f358812952f6aa2f427b235f4802d9141ed04b08cc" exitCode=0 Oct 08 14:45:52 crc kubenswrapper[4789]: I1008 14:45:52.847513 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" event={"ID":"2066f52b-bb99-4a61-98e4-63f53b8ec7c7","Type":"ContainerDied","Data":"f534a4fbf1eaecc252e5c1f358812952f6aa2f427b235f4802d9141ed04b08cc"} Oct 08 14:45:52 crc kubenswrapper[4789]: I1008 14:45:52.850633 4789 generic.go:334] "Generic (PLEG): container finished" podID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerID="f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259" exitCode=0 Oct 08 14:45:52 crc kubenswrapper[4789]: I1008 14:45:52.851153 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kz4nf" event={"ID":"bc8b8740-c44b-4eb9-aa37-4071f377d07e","Type":"ContainerDied","Data":"f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259"} Oct 08 14:45:52 crc kubenswrapper[4789]: I1008 14:45:52.851452 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kz4nf" event={"ID":"bc8b8740-c44b-4eb9-aa37-4071f377d07e","Type":"ContainerStarted","Data":"9398f66414f7015e7f9412aa0594fc2611c43cafe9bfdb3520447c6d5de7f7f8"} Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.286595 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.369661 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-489c9\" (UniqueName: \"kubernetes.io/projected/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-kube-api-access-489c9\") pod \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.369957 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-combined-ca-bundle\") pod \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.370095 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-ssh-key\") pod \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.370169 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-inventory\") pod \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.370257 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-secret-0\") pod \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\" (UID: \"2066f52b-bb99-4a61-98e4-63f53b8ec7c7\") " Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.376069 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-kube-api-access-489c9" (OuterVolumeSpecName: "kube-api-access-489c9") pod "2066f52b-bb99-4a61-98e4-63f53b8ec7c7" (UID: "2066f52b-bb99-4a61-98e4-63f53b8ec7c7"). InnerVolumeSpecName "kube-api-access-489c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.379097 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2066f52b-bb99-4a61-98e4-63f53b8ec7c7" (UID: "2066f52b-bb99-4a61-98e4-63f53b8ec7c7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.400896 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-inventory" (OuterVolumeSpecName: "inventory") pod "2066f52b-bb99-4a61-98e4-63f53b8ec7c7" (UID: "2066f52b-bb99-4a61-98e4-63f53b8ec7c7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.402084 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2066f52b-bb99-4a61-98e4-63f53b8ec7c7" (UID: "2066f52b-bb99-4a61-98e4-63f53b8ec7c7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.402481 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "2066f52b-bb99-4a61-98e4-63f53b8ec7c7" (UID: "2066f52b-bb99-4a61-98e4-63f53b8ec7c7"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.473230 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-489c9\" (UniqueName: \"kubernetes.io/projected/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-kube-api-access-489c9\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.473272 4789 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.473281 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.473290 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.473299 4789 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2066f52b-bb99-4a61-98e4-63f53b8ec7c7-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.871256 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.871234 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg" event={"ID":"2066f52b-bb99-4a61-98e4-63f53b8ec7c7","Type":"ContainerDied","Data":"156f372cb8656337666844a6a8c8379c4dba484672b7d930eefc7ed92a742839"} Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.871709 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="156f372cb8656337666844a6a8c8379c4dba484672b7d930eefc7ed92a742839" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.873421 4789 generic.go:334] "Generic (PLEG): container finished" podID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerID="71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5" exitCode=0 Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.873448 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kz4nf" event={"ID":"bc8b8740-c44b-4eb9-aa37-4071f377d07e","Type":"ContainerDied","Data":"71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5"} Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.962936 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb"] Oct 08 14:45:54 crc kubenswrapper[4789]: E1008 14:45:54.963594 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2066f52b-bb99-4a61-98e4-63f53b8ec7c7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.963620 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="2066f52b-bb99-4a61-98e4-63f53b8ec7c7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.963876 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="2066f52b-bb99-4a61-98e4-63f53b8ec7c7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.964915 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.968079 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.968463 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.968660 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.968805 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.968938 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.969204 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.969777 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:45:54 crc kubenswrapper[4789]: I1008 14:45:54.977143 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb"] Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.087628 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.087716 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.087740 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z62t\" (UniqueName: \"kubernetes.io/projected/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-kube-api-access-5z62t\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.087835 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.087913 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.088016 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.088079 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.088210 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.088353 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.189851 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.189943 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.189966 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z62t\" (UniqueName: \"kubernetes.io/projected/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-kube-api-access-5z62t\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.190008 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.190047 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.190197 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.190229 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.190614 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.191058 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.191259 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.194846 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.194850 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.194980 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.195094 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.195497 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.198545 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.199261 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.211708 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z62t\" (UniqueName: \"kubernetes.io/projected/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-kube-api-access-5z62t\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gbfjb\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.292131 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.804537 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb"] Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.884844 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" event={"ID":"b8ce1628-fa93-4910-af5a-ebce3d8ecc30","Type":"ContainerStarted","Data":"d4bbfbdd66ebbe0e00bf7f509196c896424e76834ebbdbf5248654416f74630d"} Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.887496 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kz4nf" event={"ID":"bc8b8740-c44b-4eb9-aa37-4071f377d07e","Type":"ContainerStarted","Data":"dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03"} Oct 08 14:45:55 crc kubenswrapper[4789]: I1008 14:45:55.908507 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kz4nf" podStartSLOduration=2.422002518 podStartE2EDuration="4.908484562s" podCreationTimestamp="2025-10-08 14:45:51 +0000 UTC" firstStartedPulling="2025-10-08 14:45:52.853835354 +0000 UTC m=+2692.760582846" lastFinishedPulling="2025-10-08 14:45:55.340317398 +0000 UTC m=+2695.247064890" observedRunningTime="2025-10-08 14:45:55.904323928 +0000 UTC m=+2695.811071420" watchObservedRunningTime="2025-10-08 14:45:55.908484562 +0000 UTC m=+2695.815232054" Oct 08 14:45:56 crc kubenswrapper[4789]: I1008 14:45:56.433468 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:45:56 crc kubenswrapper[4789]: I1008 14:45:56.433879 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:45:56 crc kubenswrapper[4789]: I1008 14:45:56.896949 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" event={"ID":"b8ce1628-fa93-4910-af5a-ebce3d8ecc30","Type":"ContainerStarted","Data":"86ccee050eaa72882d5541201c13298545215aee4d659c66152893298b7d7dbd"} Oct 08 14:45:56 crc kubenswrapper[4789]: I1008 14:45:56.922763 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" podStartSLOduration=2.453039104 podStartE2EDuration="2.92274162s" podCreationTimestamp="2025-10-08 14:45:54 +0000 UTC" firstStartedPulling="2025-10-08 14:45:55.808453331 +0000 UTC m=+2695.715200813" lastFinishedPulling="2025-10-08 14:45:56.278155837 +0000 UTC m=+2696.184903329" observedRunningTime="2025-10-08 14:45:56.917319373 +0000 UTC m=+2696.824066895" watchObservedRunningTime="2025-10-08 14:45:56.92274162 +0000 UTC m=+2696.829489122" Oct 08 14:46:01 crc kubenswrapper[4789]: I1008 14:46:01.799475 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:46:01 crc kubenswrapper[4789]: I1008 14:46:01.800132 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:46:01 crc kubenswrapper[4789]: I1008 14:46:01.875387 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:46:02 crc kubenswrapper[4789]: I1008 14:46:02.006707 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:46:02 crc kubenswrapper[4789]: I1008 14:46:02.114917 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kz4nf"] Oct 08 14:46:03 crc kubenswrapper[4789]: I1008 14:46:03.964597 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kz4nf" podUID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerName="registry-server" containerID="cri-o://dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03" gracePeriod=2 Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.435023 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.602393 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-utilities\") pod \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.602456 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5kf5\" (UniqueName: \"kubernetes.io/projected/bc8b8740-c44b-4eb9-aa37-4071f377d07e-kube-api-access-d5kf5\") pod \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.602526 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-catalog-content\") pod \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\" (UID: \"bc8b8740-c44b-4eb9-aa37-4071f377d07e\") " Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.604637 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-utilities" (OuterVolumeSpecName: "utilities") pod "bc8b8740-c44b-4eb9-aa37-4071f377d07e" (UID: "bc8b8740-c44b-4eb9-aa37-4071f377d07e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.608411 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc8b8740-c44b-4eb9-aa37-4071f377d07e-kube-api-access-d5kf5" (OuterVolumeSpecName: "kube-api-access-d5kf5") pod "bc8b8740-c44b-4eb9-aa37-4071f377d07e" (UID: "bc8b8740-c44b-4eb9-aa37-4071f377d07e"). InnerVolumeSpecName "kube-api-access-d5kf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.617027 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc8b8740-c44b-4eb9-aa37-4071f377d07e" (UID: "bc8b8740-c44b-4eb9-aa37-4071f377d07e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.705344 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.705395 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5kf5\" (UniqueName: \"kubernetes.io/projected/bc8b8740-c44b-4eb9-aa37-4071f377d07e-kube-api-access-d5kf5\") on node \"crc\" DevicePath \"\"" Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.705408 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc8b8740-c44b-4eb9-aa37-4071f377d07e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.974790 4789 generic.go:334] "Generic (PLEG): container finished" podID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerID="dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03" exitCode=0 Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.974831 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kz4nf" event={"ID":"bc8b8740-c44b-4eb9-aa37-4071f377d07e","Type":"ContainerDied","Data":"dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03"} Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.974863 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kz4nf" event={"ID":"bc8b8740-c44b-4eb9-aa37-4071f377d07e","Type":"ContainerDied","Data":"9398f66414f7015e7f9412aa0594fc2611c43cafe9bfdb3520447c6d5de7f7f8"} Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.974880 4789 scope.go:117] "RemoveContainer" containerID="dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03" Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.974889 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kz4nf" Oct 08 14:46:04 crc kubenswrapper[4789]: I1008 14:46:04.994715 4789 scope.go:117] "RemoveContainer" containerID="71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5" Oct 08 14:46:05 crc kubenswrapper[4789]: I1008 14:46:05.007153 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kz4nf"] Oct 08 14:46:05 crc kubenswrapper[4789]: I1008 14:46:05.015283 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kz4nf"] Oct 08 14:46:05 crc kubenswrapper[4789]: I1008 14:46:05.022131 4789 scope.go:117] "RemoveContainer" containerID="f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259" Oct 08 14:46:05 crc kubenswrapper[4789]: I1008 14:46:05.070342 4789 scope.go:117] "RemoveContainer" containerID="dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03" Oct 08 14:46:05 crc kubenswrapper[4789]: E1008 14:46:05.070791 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03\": container with ID starting with dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03 not found: ID does not exist" containerID="dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03" Oct 08 14:46:05 crc kubenswrapper[4789]: I1008 14:46:05.070823 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03"} err="failed to get container status \"dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03\": rpc error: code = NotFound desc = could not find container \"dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03\": container with ID starting with dfad7952884ff49aa00c4d9080d4ad1b52145b1e70773bbcab2c2a70bbdb2d03 not found: ID does not exist" Oct 08 14:46:05 crc kubenswrapper[4789]: I1008 14:46:05.070850 4789 scope.go:117] "RemoveContainer" containerID="71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5" Oct 08 14:46:05 crc kubenswrapper[4789]: E1008 14:46:05.071196 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5\": container with ID starting with 71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5 not found: ID does not exist" containerID="71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5" Oct 08 14:46:05 crc kubenswrapper[4789]: I1008 14:46:05.071219 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5"} err="failed to get container status \"71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5\": rpc error: code = NotFound desc = could not find container \"71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5\": container with ID starting with 71de2671649ba9dc5d7b11e30e7ccf356f90ab0342a371fcdab600178cebf5c5 not found: ID does not exist" Oct 08 14:46:05 crc kubenswrapper[4789]: I1008 14:46:05.071233 4789 scope.go:117] "RemoveContainer" containerID="f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259" Oct 08 14:46:05 crc kubenswrapper[4789]: E1008 14:46:05.071531 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259\": container with ID starting with f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259 not found: ID does not exist" containerID="f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259" Oct 08 14:46:05 crc kubenswrapper[4789]: I1008 14:46:05.071573 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259"} err="failed to get container status \"f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259\": rpc error: code = NotFound desc = could not find container \"f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259\": container with ID starting with f46b5daf9e210a56862003951657837f9e477b2c4de23845308661d665fc5259 not found: ID does not exist" Oct 08 14:46:06 crc kubenswrapper[4789]: I1008 14:46:06.741502 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" path="/var/lib/kubelet/pods/bc8b8740-c44b-4eb9-aa37-4071f377d07e/volumes" Oct 08 14:46:26 crc kubenswrapper[4789]: I1008 14:46:26.433113 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:46:26 crc kubenswrapper[4789]: I1008 14:46:26.433650 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:46:56 crc kubenswrapper[4789]: I1008 14:46:56.433095 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:46:56 crc kubenswrapper[4789]: I1008 14:46:56.433783 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:46:56 crc kubenswrapper[4789]: I1008 14:46:56.433853 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:46:56 crc kubenswrapper[4789]: I1008 14:46:56.435369 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5f903459af49810b12480e53c24b2593b65c7aac101fb02e8a672f71a5607dbb"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:46:56 crc kubenswrapper[4789]: I1008 14:46:56.435516 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://5f903459af49810b12480e53c24b2593b65c7aac101fb02e8a672f71a5607dbb" gracePeriod=600 Oct 08 14:46:57 crc kubenswrapper[4789]: I1008 14:46:57.492385 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="5f903459af49810b12480e53c24b2593b65c7aac101fb02e8a672f71a5607dbb" exitCode=0 Oct 08 14:46:57 crc kubenswrapper[4789]: I1008 14:46:57.492451 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"5f903459af49810b12480e53c24b2593b65c7aac101fb02e8a672f71a5607dbb"} Oct 08 14:46:57 crc kubenswrapper[4789]: I1008 14:46:57.495115 4789 scope.go:117] "RemoveContainer" containerID="2f7fa2b6ac2c2dbea782f3d44deb4334dd2183e8b7d4504ded8973d1300cbe6b" Oct 08 14:46:58 crc kubenswrapper[4789]: I1008 14:46:58.514269 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170"} Oct 08 14:49:17 crc kubenswrapper[4789]: I1008 14:49:17.841365 4789 generic.go:334] "Generic (PLEG): container finished" podID="b8ce1628-fa93-4910-af5a-ebce3d8ecc30" containerID="86ccee050eaa72882d5541201c13298545215aee4d659c66152893298b7d7dbd" exitCode=0 Oct 08 14:49:17 crc kubenswrapper[4789]: I1008 14:49:17.841494 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" event={"ID":"b8ce1628-fa93-4910-af5a-ebce3d8ecc30","Type":"ContainerDied","Data":"86ccee050eaa72882d5541201c13298545215aee4d659c66152893298b7d7dbd"} Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.279693 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.369271 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-0\") pod \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.369933 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-inventory\") pod \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.370221 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-combined-ca-bundle\") pod \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.370431 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-ssh-key\") pod \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.370745 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-1\") pod \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.371309 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z62t\" (UniqueName: \"kubernetes.io/projected/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-kube-api-access-5z62t\") pod \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.371547 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-0\") pod \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.371801 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-1\") pod \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.374380 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-extra-config-0\") pod \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\" (UID: \"b8ce1628-fa93-4910-af5a-ebce3d8ecc30\") " Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.377177 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b8ce1628-fa93-4910-af5a-ebce3d8ecc30" (UID: "b8ce1628-fa93-4910-af5a-ebce3d8ecc30"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.377690 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-kube-api-access-5z62t" (OuterVolumeSpecName: "kube-api-access-5z62t") pod "b8ce1628-fa93-4910-af5a-ebce3d8ecc30" (UID: "b8ce1628-fa93-4910-af5a-ebce3d8ecc30"). InnerVolumeSpecName "kube-api-access-5z62t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.397349 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "b8ce1628-fa93-4910-af5a-ebce3d8ecc30" (UID: "b8ce1628-fa93-4910-af5a-ebce3d8ecc30"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.399487 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-inventory" (OuterVolumeSpecName: "inventory") pod "b8ce1628-fa93-4910-af5a-ebce3d8ecc30" (UID: "b8ce1628-fa93-4910-af5a-ebce3d8ecc30"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.407750 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "b8ce1628-fa93-4910-af5a-ebce3d8ecc30" (UID: "b8ce1628-fa93-4910-af5a-ebce3d8ecc30"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.409674 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "b8ce1628-fa93-4910-af5a-ebce3d8ecc30" (UID: "b8ce1628-fa93-4910-af5a-ebce3d8ecc30"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.410590 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8ce1628-fa93-4910-af5a-ebce3d8ecc30" (UID: "b8ce1628-fa93-4910-af5a-ebce3d8ecc30"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.416792 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "b8ce1628-fa93-4910-af5a-ebce3d8ecc30" (UID: "b8ce1628-fa93-4910-af5a-ebce3d8ecc30"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.421696 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "b8ce1628-fa93-4910-af5a-ebce3d8ecc30" (UID: "b8ce1628-fa93-4910-af5a-ebce3d8ecc30"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.478367 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z62t\" (UniqueName: \"kubernetes.io/projected/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-kube-api-access-5z62t\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.478419 4789 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.478434 4789 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.478444 4789 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.478452 4789 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.478462 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.478473 4789 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.478481 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.478491 4789 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b8ce1628-fa93-4910-af5a-ebce3d8ecc30-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.859870 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" event={"ID":"b8ce1628-fa93-4910-af5a-ebce3d8ecc30","Type":"ContainerDied","Data":"d4bbfbdd66ebbe0e00bf7f509196c896424e76834ebbdbf5248654416f74630d"} Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.859916 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4bbfbdd66ebbe0e00bf7f509196c896424e76834ebbdbf5248654416f74630d" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.859944 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gbfjb" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.974612 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk"] Oct 08 14:49:19 crc kubenswrapper[4789]: E1008 14:49:19.976743 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ce1628-fa93-4910-af5a-ebce3d8ecc30" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.976824 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ce1628-fa93-4910-af5a-ebce3d8ecc30" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 08 14:49:19 crc kubenswrapper[4789]: E1008 14:49:19.976892 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerName="extract-content" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.976942 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerName="extract-content" Oct 08 14:49:19 crc kubenswrapper[4789]: E1008 14:49:19.977087 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerName="registry-server" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.977152 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerName="registry-server" Oct 08 14:49:19 crc kubenswrapper[4789]: E1008 14:49:19.977220 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerName="extract-utilities" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.977279 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerName="extract-utilities" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.977506 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ce1628-fa93-4910-af5a-ebce3d8ecc30" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.977588 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc8b8740-c44b-4eb9-aa37-4071f377d07e" containerName="registry-server" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.978269 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.981863 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.981931 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.981940 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.983746 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 14:49:19 crc kubenswrapper[4789]: I1008 14:49:19.983942 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tkx2n" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.022649 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk"] Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.127724 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.127794 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.127913 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.128057 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.128138 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.128184 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.128223 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcpd8\" (UniqueName: \"kubernetes.io/projected/c9031f68-f885-456b-a1ea-ce7742e0eccc-kube-api-access-rcpd8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.229804 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcpd8\" (UniqueName: \"kubernetes.io/projected/c9031f68-f885-456b-a1ea-ce7742e0eccc-kube-api-access-rcpd8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.229874 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.229919 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.230029 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.230174 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.230237 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.230289 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.235085 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.235107 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.235692 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.235947 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.236564 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.238291 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.251818 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcpd8\" (UniqueName: \"kubernetes.io/projected/c9031f68-f885-456b-a1ea-ce7742e0eccc-kube-api-access-rcpd8\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.318221 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:49:20 crc kubenswrapper[4789]: I1008 14:49:20.863347 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk"] Oct 08 14:49:21 crc kubenswrapper[4789]: I1008 14:49:21.877073 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" event={"ID":"c9031f68-f885-456b-a1ea-ce7742e0eccc","Type":"ContainerStarted","Data":"c763b00506b4e43959eaac2ce15fc2015d269e43338ef3e8757260e80313031d"} Oct 08 14:49:21 crc kubenswrapper[4789]: I1008 14:49:21.877326 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" event={"ID":"c9031f68-f885-456b-a1ea-ce7742e0eccc","Type":"ContainerStarted","Data":"0b472609d7f26c1718ff339d1d771d563c301e9fb59e443e95997c8a66a64797"} Oct 08 14:49:26 crc kubenswrapper[4789]: I1008 14:49:26.432492 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:49:26 crc kubenswrapper[4789]: I1008 14:49:26.432973 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:49:26 crc kubenswrapper[4789]: I1008 14:49:26.865371 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" podStartSLOduration=7.227531349 podStartE2EDuration="7.865348179s" podCreationTimestamp="2025-10-08 14:49:19 +0000 UTC" firstStartedPulling="2025-10-08 14:49:20.871541784 +0000 UTC m=+2900.778289276" lastFinishedPulling="2025-10-08 14:49:21.509358614 +0000 UTC m=+2901.416106106" observedRunningTime="2025-10-08 14:49:21.902196289 +0000 UTC m=+2901.808943781" watchObservedRunningTime="2025-10-08 14:49:26.865348179 +0000 UTC m=+2906.772095671" Oct 08 14:49:26 crc kubenswrapper[4789]: I1008 14:49:26.867408 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ttnwx"] Oct 08 14:49:26 crc kubenswrapper[4789]: I1008 14:49:26.869604 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:26 crc kubenswrapper[4789]: I1008 14:49:26.880051 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ttnwx"] Oct 08 14:49:26 crc kubenswrapper[4789]: I1008 14:49:26.961062 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n255r\" (UniqueName: \"kubernetes.io/projected/6320ad5a-e7c9-4308-a2ce-56087abb3414-kube-api-access-n255r\") pod \"certified-operators-ttnwx\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:26 crc kubenswrapper[4789]: I1008 14:49:26.961140 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-catalog-content\") pod \"certified-operators-ttnwx\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:26 crc kubenswrapper[4789]: I1008 14:49:26.961285 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-utilities\") pod \"certified-operators-ttnwx\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:27 crc kubenswrapper[4789]: I1008 14:49:27.063548 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n255r\" (UniqueName: \"kubernetes.io/projected/6320ad5a-e7c9-4308-a2ce-56087abb3414-kube-api-access-n255r\") pod \"certified-operators-ttnwx\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:27 crc kubenswrapper[4789]: I1008 14:49:27.063614 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-catalog-content\") pod \"certified-operators-ttnwx\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:27 crc kubenswrapper[4789]: I1008 14:49:27.063719 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-utilities\") pod \"certified-operators-ttnwx\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:27 crc kubenswrapper[4789]: I1008 14:49:27.064184 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-catalog-content\") pod \"certified-operators-ttnwx\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:27 crc kubenswrapper[4789]: I1008 14:49:27.064204 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-utilities\") pod \"certified-operators-ttnwx\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:27 crc kubenswrapper[4789]: I1008 14:49:27.082107 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n255r\" (UniqueName: \"kubernetes.io/projected/6320ad5a-e7c9-4308-a2ce-56087abb3414-kube-api-access-n255r\") pod \"certified-operators-ttnwx\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:27 crc kubenswrapper[4789]: I1008 14:49:27.200240 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:27 crc kubenswrapper[4789]: I1008 14:49:27.706798 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ttnwx"] Oct 08 14:49:27 crc kubenswrapper[4789]: I1008 14:49:27.948817 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttnwx" event={"ID":"6320ad5a-e7c9-4308-a2ce-56087abb3414","Type":"ContainerStarted","Data":"aa2072d30b5b63a91895f18af8fd2bfa32a73ba920d9ff5a203e79d456a65eac"} Oct 08 14:49:28 crc kubenswrapper[4789]: I1008 14:49:28.963735 4789 generic.go:334] "Generic (PLEG): container finished" podID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerID="767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2" exitCode=0 Oct 08 14:49:28 crc kubenswrapper[4789]: I1008 14:49:28.964008 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttnwx" event={"ID":"6320ad5a-e7c9-4308-a2ce-56087abb3414","Type":"ContainerDied","Data":"767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2"} Oct 08 14:49:30 crc kubenswrapper[4789]: I1008 14:49:30.983537 4789 generic.go:334] "Generic (PLEG): container finished" podID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerID="a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719" exitCode=0 Oct 08 14:49:30 crc kubenswrapper[4789]: I1008 14:49:30.983641 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttnwx" event={"ID":"6320ad5a-e7c9-4308-a2ce-56087abb3414","Type":"ContainerDied","Data":"a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719"} Oct 08 14:49:31 crc kubenswrapper[4789]: I1008 14:49:31.993781 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttnwx" event={"ID":"6320ad5a-e7c9-4308-a2ce-56087abb3414","Type":"ContainerStarted","Data":"c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3"} Oct 08 14:49:32 crc kubenswrapper[4789]: I1008 14:49:32.016781 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ttnwx" podStartSLOduration=3.284248325 podStartE2EDuration="6.01676205s" podCreationTimestamp="2025-10-08 14:49:26 +0000 UTC" firstStartedPulling="2025-10-08 14:49:28.966332117 +0000 UTC m=+2908.873079609" lastFinishedPulling="2025-10-08 14:49:31.698845842 +0000 UTC m=+2911.605593334" observedRunningTime="2025-10-08 14:49:32.009484052 +0000 UTC m=+2911.916231544" watchObservedRunningTime="2025-10-08 14:49:32.01676205 +0000 UTC m=+2911.923509542" Oct 08 14:49:37 crc kubenswrapper[4789]: I1008 14:49:37.201080 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:37 crc kubenswrapper[4789]: I1008 14:49:37.202941 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:37 crc kubenswrapper[4789]: I1008 14:49:37.250493 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:38 crc kubenswrapper[4789]: I1008 14:49:38.094317 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:38 crc kubenswrapper[4789]: I1008 14:49:38.145671 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ttnwx"] Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.061608 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ttnwx" podUID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerName="registry-server" containerID="cri-o://c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3" gracePeriod=2 Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.514475 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.644584 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-utilities\") pod \"6320ad5a-e7c9-4308-a2ce-56087abb3414\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.644713 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n255r\" (UniqueName: \"kubernetes.io/projected/6320ad5a-e7c9-4308-a2ce-56087abb3414-kube-api-access-n255r\") pod \"6320ad5a-e7c9-4308-a2ce-56087abb3414\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.644732 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-catalog-content\") pod \"6320ad5a-e7c9-4308-a2ce-56087abb3414\" (UID: \"6320ad5a-e7c9-4308-a2ce-56087abb3414\") " Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.645933 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-utilities" (OuterVolumeSpecName: "utilities") pod "6320ad5a-e7c9-4308-a2ce-56087abb3414" (UID: "6320ad5a-e7c9-4308-a2ce-56087abb3414"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.654842 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6320ad5a-e7c9-4308-a2ce-56087abb3414-kube-api-access-n255r" (OuterVolumeSpecName: "kube-api-access-n255r") pod "6320ad5a-e7c9-4308-a2ce-56087abb3414" (UID: "6320ad5a-e7c9-4308-a2ce-56087abb3414"). InnerVolumeSpecName "kube-api-access-n255r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.700603 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6320ad5a-e7c9-4308-a2ce-56087abb3414" (UID: "6320ad5a-e7c9-4308-a2ce-56087abb3414"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.747107 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.747337 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n255r\" (UniqueName: \"kubernetes.io/projected/6320ad5a-e7c9-4308-a2ce-56087abb3414-kube-api-access-n255r\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:40 crc kubenswrapper[4789]: I1008 14:49:40.747427 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6320ad5a-e7c9-4308-a2ce-56087abb3414-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.073781 4789 generic.go:334] "Generic (PLEG): container finished" podID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerID="c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3" exitCode=0 Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.073893 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttnwx" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.073881 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttnwx" event={"ID":"6320ad5a-e7c9-4308-a2ce-56087abb3414","Type":"ContainerDied","Data":"c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3"} Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.074364 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttnwx" event={"ID":"6320ad5a-e7c9-4308-a2ce-56087abb3414","Type":"ContainerDied","Data":"aa2072d30b5b63a91895f18af8fd2bfa32a73ba920d9ff5a203e79d456a65eac"} Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.074396 4789 scope.go:117] "RemoveContainer" containerID="c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.102636 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ttnwx"] Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.118545 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ttnwx"] Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.120956 4789 scope.go:117] "RemoveContainer" containerID="a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.150663 4789 scope.go:117] "RemoveContainer" containerID="767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.192883 4789 scope.go:117] "RemoveContainer" containerID="c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3" Oct 08 14:49:41 crc kubenswrapper[4789]: E1008 14:49:41.193473 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3\": container with ID starting with c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3 not found: ID does not exist" containerID="c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.193522 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3"} err="failed to get container status \"c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3\": rpc error: code = NotFound desc = could not find container \"c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3\": container with ID starting with c3b27df26a68b5cc28e3833914dc1487e943ddb96b7e4819112f1d1e18ab87e3 not found: ID does not exist" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.193553 4789 scope.go:117] "RemoveContainer" containerID="a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719" Oct 08 14:49:41 crc kubenswrapper[4789]: E1008 14:49:41.194131 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719\": container with ID starting with a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719 not found: ID does not exist" containerID="a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.194188 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719"} err="failed to get container status \"a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719\": rpc error: code = NotFound desc = could not find container \"a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719\": container with ID starting with a431d7ef7f91187d411038c54d64614287a1a7e251100080a6de42dc8ce71719 not found: ID does not exist" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.194223 4789 scope.go:117] "RemoveContainer" containerID="767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2" Oct 08 14:49:41 crc kubenswrapper[4789]: E1008 14:49:41.194586 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2\": container with ID starting with 767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2 not found: ID does not exist" containerID="767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2" Oct 08 14:49:41 crc kubenswrapper[4789]: I1008 14:49:41.194621 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2"} err="failed to get container status \"767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2\": rpc error: code = NotFound desc = could not find container \"767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2\": container with ID starting with 767344c7b771d436408d944ac78f92b854c040ef7479359c20524e27271a59e2 not found: ID does not exist" Oct 08 14:49:42 crc kubenswrapper[4789]: I1008 14:49:42.740420 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6320ad5a-e7c9-4308-a2ce-56087abb3414" path="/var/lib/kubelet/pods/6320ad5a-e7c9-4308-a2ce-56087abb3414/volumes" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.534382 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vzcpp"] Oct 08 14:49:43 crc kubenswrapper[4789]: E1008 14:49:43.534819 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerName="registry-server" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.534841 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerName="registry-server" Oct 08 14:49:43 crc kubenswrapper[4789]: E1008 14:49:43.534854 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerName="extract-utilities" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.534860 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerName="extract-utilities" Oct 08 14:49:43 crc kubenswrapper[4789]: E1008 14:49:43.534878 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerName="extract-content" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.534884 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerName="extract-content" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.535104 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6320ad5a-e7c9-4308-a2ce-56087abb3414" containerName="registry-server" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.536646 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.561220 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vzcpp"] Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.605333 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-catalog-content\") pod \"redhat-operators-vzcpp\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.605469 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvzn5\" (UniqueName: \"kubernetes.io/projected/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-kube-api-access-pvzn5\") pod \"redhat-operators-vzcpp\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.605553 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-utilities\") pod \"redhat-operators-vzcpp\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.707151 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-catalog-content\") pod \"redhat-operators-vzcpp\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.707231 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvzn5\" (UniqueName: \"kubernetes.io/projected/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-kube-api-access-pvzn5\") pod \"redhat-operators-vzcpp\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.707300 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-utilities\") pod \"redhat-operators-vzcpp\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.707732 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-catalog-content\") pod \"redhat-operators-vzcpp\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.707798 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-utilities\") pod \"redhat-operators-vzcpp\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.730049 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvzn5\" (UniqueName: \"kubernetes.io/projected/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-kube-api-access-pvzn5\") pod \"redhat-operators-vzcpp\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:43 crc kubenswrapper[4789]: I1008 14:49:43.860760 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:44 crc kubenswrapper[4789]: W1008 14:49:44.320136 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7565353a_92b8_455c_bd6c_5e2e1fbaa72f.slice/crio-42404440636ebd7b36e3f45f227b32272f082db6c01e5f291a7cf03db51e20e0 WatchSource:0}: Error finding container 42404440636ebd7b36e3f45f227b32272f082db6c01e5f291a7cf03db51e20e0: Status 404 returned error can't find the container with id 42404440636ebd7b36e3f45f227b32272f082db6c01e5f291a7cf03db51e20e0 Oct 08 14:49:44 crc kubenswrapper[4789]: I1008 14:49:44.321433 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vzcpp"] Oct 08 14:49:45 crc kubenswrapper[4789]: I1008 14:49:45.112424 4789 generic.go:334] "Generic (PLEG): container finished" podID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerID="1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b" exitCode=0 Oct 08 14:49:45 crc kubenswrapper[4789]: I1008 14:49:45.112547 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcpp" event={"ID":"7565353a-92b8-455c-bd6c-5e2e1fbaa72f","Type":"ContainerDied","Data":"1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b"} Oct 08 14:49:45 crc kubenswrapper[4789]: I1008 14:49:45.114470 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcpp" event={"ID":"7565353a-92b8-455c-bd6c-5e2e1fbaa72f","Type":"ContainerStarted","Data":"42404440636ebd7b36e3f45f227b32272f082db6c01e5f291a7cf03db51e20e0"} Oct 08 14:49:46 crc kubenswrapper[4789]: I1008 14:49:46.137774 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcpp" event={"ID":"7565353a-92b8-455c-bd6c-5e2e1fbaa72f","Type":"ContainerStarted","Data":"1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a"} Oct 08 14:49:50 crc kubenswrapper[4789]: I1008 14:49:50.185634 4789 generic.go:334] "Generic (PLEG): container finished" podID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerID="1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a" exitCode=0 Oct 08 14:49:50 crc kubenswrapper[4789]: I1008 14:49:50.185744 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcpp" event={"ID":"7565353a-92b8-455c-bd6c-5e2e1fbaa72f","Type":"ContainerDied","Data":"1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a"} Oct 08 14:49:51 crc kubenswrapper[4789]: I1008 14:49:51.197922 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcpp" event={"ID":"7565353a-92b8-455c-bd6c-5e2e1fbaa72f","Type":"ContainerStarted","Data":"8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33"} Oct 08 14:49:51 crc kubenswrapper[4789]: I1008 14:49:51.219050 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vzcpp" podStartSLOduration=2.43113842 podStartE2EDuration="8.219035944s" podCreationTimestamp="2025-10-08 14:49:43 +0000 UTC" firstStartedPulling="2025-10-08 14:49:45.114078335 +0000 UTC m=+2925.020825847" lastFinishedPulling="2025-10-08 14:49:50.901975879 +0000 UTC m=+2930.808723371" observedRunningTime="2025-10-08 14:49:51.217509732 +0000 UTC m=+2931.124257224" watchObservedRunningTime="2025-10-08 14:49:51.219035944 +0000 UTC m=+2931.125783436" Oct 08 14:49:53 crc kubenswrapper[4789]: I1008 14:49:53.861069 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:53 crc kubenswrapper[4789]: I1008 14:49:53.861311 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:49:54 crc kubenswrapper[4789]: I1008 14:49:54.904643 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vzcpp" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerName="registry-server" probeResult="failure" output=< Oct 08 14:49:54 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 14:49:54 crc kubenswrapper[4789]: > Oct 08 14:49:56 crc kubenswrapper[4789]: I1008 14:49:56.433790 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:49:56 crc kubenswrapper[4789]: I1008 14:49:56.434187 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:50:03 crc kubenswrapper[4789]: I1008 14:50:03.921980 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:50:03 crc kubenswrapper[4789]: I1008 14:50:03.986565 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:50:04 crc kubenswrapper[4789]: I1008 14:50:04.159199 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vzcpp"] Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.335120 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vzcpp" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerName="registry-server" containerID="cri-o://8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33" gracePeriod=2 Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.760135 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.848498 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvzn5\" (UniqueName: \"kubernetes.io/projected/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-kube-api-access-pvzn5\") pod \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.848567 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-catalog-content\") pod \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.849466 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-utilities\") pod \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\" (UID: \"7565353a-92b8-455c-bd6c-5e2e1fbaa72f\") " Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.850791 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-utilities" (OuterVolumeSpecName: "utilities") pod "7565353a-92b8-455c-bd6c-5e2e1fbaa72f" (UID: "7565353a-92b8-455c-bd6c-5e2e1fbaa72f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.855452 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-kube-api-access-pvzn5" (OuterVolumeSpecName: "kube-api-access-pvzn5") pod "7565353a-92b8-455c-bd6c-5e2e1fbaa72f" (UID: "7565353a-92b8-455c-bd6c-5e2e1fbaa72f"). InnerVolumeSpecName "kube-api-access-pvzn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.936031 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7565353a-92b8-455c-bd6c-5e2e1fbaa72f" (UID: "7565353a-92b8-455c-bd6c-5e2e1fbaa72f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.951568 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvzn5\" (UniqueName: \"kubernetes.io/projected/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-kube-api-access-pvzn5\") on node \"crc\" DevicePath \"\"" Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.951599 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:50:05 crc kubenswrapper[4789]: I1008 14:50:05.951609 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7565353a-92b8-455c-bd6c-5e2e1fbaa72f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.347431 4789 generic.go:334] "Generic (PLEG): container finished" podID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerID="8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33" exitCode=0 Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.347479 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcpp" event={"ID":"7565353a-92b8-455c-bd6c-5e2e1fbaa72f","Type":"ContainerDied","Data":"8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33"} Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.347514 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcpp" event={"ID":"7565353a-92b8-455c-bd6c-5e2e1fbaa72f","Type":"ContainerDied","Data":"42404440636ebd7b36e3f45f227b32272f082db6c01e5f291a7cf03db51e20e0"} Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.347516 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzcpp" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.347533 4789 scope.go:117] "RemoveContainer" containerID="8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.378053 4789 scope.go:117] "RemoveContainer" containerID="1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.388084 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vzcpp"] Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.402368 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vzcpp"] Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.416440 4789 scope.go:117] "RemoveContainer" containerID="1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.472174 4789 scope.go:117] "RemoveContainer" containerID="8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33" Oct 08 14:50:06 crc kubenswrapper[4789]: E1008 14:50:06.472764 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33\": container with ID starting with 8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33 not found: ID does not exist" containerID="8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.472852 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33"} err="failed to get container status \"8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33\": rpc error: code = NotFound desc = could not find container \"8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33\": container with ID starting with 8353d6aea7249d834ec61ff826bc8d64573f9cb3efe1a110435d094ada976b33 not found: ID does not exist" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.472887 4789 scope.go:117] "RemoveContainer" containerID="1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a" Oct 08 14:50:06 crc kubenswrapper[4789]: E1008 14:50:06.473367 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a\": container with ID starting with 1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a not found: ID does not exist" containerID="1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.473396 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a"} err="failed to get container status \"1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a\": rpc error: code = NotFound desc = could not find container \"1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a\": container with ID starting with 1307fc8a0fcf40dadc8a64961d528197cf270bfd5cba07caa166a7c5c5ac737a not found: ID does not exist" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.473417 4789 scope.go:117] "RemoveContainer" containerID="1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b" Oct 08 14:50:06 crc kubenswrapper[4789]: E1008 14:50:06.474160 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b\": container with ID starting with 1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b not found: ID does not exist" containerID="1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.474189 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b"} err="failed to get container status \"1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b\": rpc error: code = NotFound desc = could not find container \"1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b\": container with ID starting with 1b75fe8e7c338361aeba7b67b82cbf8fec5fa81ebe8b6041a87fbcf1e78ef75b not found: ID does not exist" Oct 08 14:50:06 crc kubenswrapper[4789]: I1008 14:50:06.741538 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" path="/var/lib/kubelet/pods/7565353a-92b8-455c-bd6c-5e2e1fbaa72f/volumes" Oct 08 14:50:26 crc kubenswrapper[4789]: I1008 14:50:26.433014 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:50:26 crc kubenswrapper[4789]: I1008 14:50:26.433511 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:50:26 crc kubenswrapper[4789]: I1008 14:50:26.433592 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:50:26 crc kubenswrapper[4789]: I1008 14:50:26.434527 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:50:26 crc kubenswrapper[4789]: I1008 14:50:26.434618 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" gracePeriod=600 Oct 08 14:50:26 crc kubenswrapper[4789]: E1008 14:50:26.558830 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:50:27 crc kubenswrapper[4789]: I1008 14:50:27.554454 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" exitCode=0 Oct 08 14:50:27 crc kubenswrapper[4789]: I1008 14:50:27.554537 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170"} Oct 08 14:50:27 crc kubenswrapper[4789]: I1008 14:50:27.554794 4789 scope.go:117] "RemoveContainer" containerID="5f903459af49810b12480e53c24b2593b65c7aac101fb02e8a672f71a5607dbb" Oct 08 14:50:27 crc kubenswrapper[4789]: I1008 14:50:27.555519 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:50:27 crc kubenswrapper[4789]: E1008 14:50:27.555940 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:50:40 crc kubenswrapper[4789]: I1008 14:50:40.752319 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:50:40 crc kubenswrapper[4789]: E1008 14:50:40.753171 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:50:52 crc kubenswrapper[4789]: I1008 14:50:52.730346 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:50:52 crc kubenswrapper[4789]: E1008 14:50:52.731076 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:51:03 crc kubenswrapper[4789]: I1008 14:51:03.731337 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:51:03 crc kubenswrapper[4789]: E1008 14:51:03.732196 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:51:15 crc kubenswrapper[4789]: I1008 14:51:15.730820 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:51:15 crc kubenswrapper[4789]: E1008 14:51:15.731688 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:51:30 crc kubenswrapper[4789]: I1008 14:51:30.737325 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:51:30 crc kubenswrapper[4789]: E1008 14:51:30.738330 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:51:36 crc kubenswrapper[4789]: I1008 14:51:36.249543 4789 generic.go:334] "Generic (PLEG): container finished" podID="c9031f68-f885-456b-a1ea-ce7742e0eccc" containerID="c763b00506b4e43959eaac2ce15fc2015d269e43338ef3e8757260e80313031d" exitCode=0 Oct 08 14:51:36 crc kubenswrapper[4789]: I1008 14:51:36.249760 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" event={"ID":"c9031f68-f885-456b-a1ea-ce7742e0eccc","Type":"ContainerDied","Data":"c763b00506b4e43959eaac2ce15fc2015d269e43338ef3e8757260e80313031d"} Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.649583 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.704837 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-inventory\") pod \"c9031f68-f885-456b-a1ea-ce7742e0eccc\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.704912 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcpd8\" (UniqueName: \"kubernetes.io/projected/c9031f68-f885-456b-a1ea-ce7742e0eccc-kube-api-access-rcpd8\") pod \"c9031f68-f885-456b-a1ea-ce7742e0eccc\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.704939 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-telemetry-combined-ca-bundle\") pod \"c9031f68-f885-456b-a1ea-ce7742e0eccc\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.704979 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-1\") pod \"c9031f68-f885-456b-a1ea-ce7742e0eccc\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.705646 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ssh-key\") pod \"c9031f68-f885-456b-a1ea-ce7742e0eccc\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.705696 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-2\") pod \"c9031f68-f885-456b-a1ea-ce7742e0eccc\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.705732 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-0\") pod \"c9031f68-f885-456b-a1ea-ce7742e0eccc\" (UID: \"c9031f68-f885-456b-a1ea-ce7742e0eccc\") " Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.711236 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c9031f68-f885-456b-a1ea-ce7742e0eccc" (UID: "c9031f68-f885-456b-a1ea-ce7742e0eccc"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.716153 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9031f68-f885-456b-a1ea-ce7742e0eccc-kube-api-access-rcpd8" (OuterVolumeSpecName: "kube-api-access-rcpd8") pod "c9031f68-f885-456b-a1ea-ce7742e0eccc" (UID: "c9031f68-f885-456b-a1ea-ce7742e0eccc"). InnerVolumeSpecName "kube-api-access-rcpd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.742288 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c9031f68-f885-456b-a1ea-ce7742e0eccc" (UID: "c9031f68-f885-456b-a1ea-ce7742e0eccc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.748768 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-inventory" (OuterVolumeSpecName: "inventory") pod "c9031f68-f885-456b-a1ea-ce7742e0eccc" (UID: "c9031f68-f885-456b-a1ea-ce7742e0eccc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.753792 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "c9031f68-f885-456b-a1ea-ce7742e0eccc" (UID: "c9031f68-f885-456b-a1ea-ce7742e0eccc"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.756068 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "c9031f68-f885-456b-a1ea-ce7742e0eccc" (UID: "c9031f68-f885-456b-a1ea-ce7742e0eccc"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.767408 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "c9031f68-f885-456b-a1ea-ce7742e0eccc" (UID: "c9031f68-f885-456b-a1ea-ce7742e0eccc"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.807612 4789 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.807643 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcpd8\" (UniqueName: \"kubernetes.io/projected/c9031f68-f885-456b-a1ea-ce7742e0eccc-kube-api-access-rcpd8\") on node \"crc\" DevicePath \"\"" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.807655 4789 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.807664 4789 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.807673 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.807682 4789 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 08 14:51:37 crc kubenswrapper[4789]: I1008 14:51:37.807691 4789 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9031f68-f885-456b-a1ea-ce7742e0eccc-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:51:38 crc kubenswrapper[4789]: I1008 14:51:38.275710 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" event={"ID":"c9031f68-f885-456b-a1ea-ce7742e0eccc","Type":"ContainerDied","Data":"0b472609d7f26c1718ff339d1d771d563c301e9fb59e443e95997c8a66a64797"} Oct 08 14:51:38 crc kubenswrapper[4789]: I1008 14:51:38.275756 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b472609d7f26c1718ff339d1d771d563c301e9fb59e443e95997c8a66a64797" Oct 08 14:51:38 crc kubenswrapper[4789]: I1008 14:51:38.275787 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk" Oct 08 14:51:41 crc kubenswrapper[4789]: I1008 14:51:41.730160 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:51:41 crc kubenswrapper[4789]: E1008 14:51:41.732359 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:51:56 crc kubenswrapper[4789]: I1008 14:51:56.730419 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:51:56 crc kubenswrapper[4789]: E1008 14:51:56.731240 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:52:11 crc kubenswrapper[4789]: I1008 14:52:11.730146 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:52:11 crc kubenswrapper[4789]: E1008 14:52:11.731056 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.463879 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 08 14:52:12 crc kubenswrapper[4789]: E1008 14:52:12.464721 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerName="registry-server" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.464744 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerName="registry-server" Oct 08 14:52:12 crc kubenswrapper[4789]: E1008 14:52:12.464765 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9031f68-f885-456b-a1ea-ce7742e0eccc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.464774 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9031f68-f885-456b-a1ea-ce7742e0eccc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 08 14:52:12 crc kubenswrapper[4789]: E1008 14:52:12.464789 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerName="extract-utilities" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.464797 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerName="extract-utilities" Oct 08 14:52:12 crc kubenswrapper[4789]: E1008 14:52:12.464808 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerName="extract-content" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.464817 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerName="extract-content" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.465099 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9031f68-f885-456b-a1ea-ce7742e0eccc" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.465133 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7565353a-92b8-455c-bd6c-5e2e1fbaa72f" containerName="registry-server" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.466454 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.468300 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.502145 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.551682 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-0"] Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554206 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-config-data\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554264 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-sys\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554327 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n28g\" (UniqueName: \"kubernetes.io/projected/6b3acb2d-0c53-40a4-8857-8e07cf54c530-kube-api-access-6n28g\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554384 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554495 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-scripts\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554526 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554577 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-lib-modules\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554606 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554627 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554707 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554750 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-run\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554769 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-dev\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554788 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554806 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-etc-nvme\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.554832 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-config-data-custom\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.557958 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.560270 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-config-data" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.564173 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.619588 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.621315 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.623821 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-2-config-data" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.631914 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656563 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-run\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656608 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-dev\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656634 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656658 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-etc-nvme\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656679 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-config-data-custom\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656703 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656733 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656750 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656776 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-config-data\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656797 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-sys\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656824 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656840 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n28g\" (UniqueName: \"kubernetes.io/projected/6b3acb2d-0c53-40a4-8857-8e07cf54c530-kube-api-access-6n28g\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656865 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656895 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656918 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656946 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.656966 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657043 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-dev\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657064 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657085 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657108 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-scripts\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657124 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657143 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657167 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657199 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-sys\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657218 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-lib-modules\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657246 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-sys\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657271 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh57g\" (UniqueName: \"kubernetes.io/projected/7b5501af-92e0-4c4a-badd-05089ae1221b-kube-api-access-lh57g\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657283 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657300 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657344 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-lib-modules\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657463 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657522 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657551 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-dev\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657552 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-run\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.657968 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.658030 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-run\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.658103 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-etc-nvme\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.658324 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.658369 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6b3acb2d-0c53-40a4-8857-8e07cf54c530-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.662645 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-scripts\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.663588 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.664181 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-config-data-custom\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.677655 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n28g\" (UniqueName: \"kubernetes.io/projected/6b3acb2d-0c53-40a4-8857-8e07cf54c530-kube-api-access-6n28g\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.685070 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3acb2d-0c53-40a4-8857-8e07cf54c530-config-data\") pod \"cinder-backup-0\" (UID: \"6b3acb2d-0c53-40a4-8857-8e07cf54c530\") " pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761229 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761345 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761393 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761424 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761469 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761497 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761522 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761556 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761587 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761625 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761652 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761691 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761719 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761755 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761796 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761841 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761879 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwws5\" (UniqueName: \"kubernetes.io/projected/e0006437-3cf9-437c-a3bd-a12f6ef84aae-kube-api-access-bwws5\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761908 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761933 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761955 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.761998 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-dev\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762032 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762065 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762096 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762131 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762173 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762225 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-sys\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762254 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh57g\" (UniqueName: \"kubernetes.io/projected/7b5501af-92e0-4c4a-badd-05089ae1221b-kube-api-access-lh57g\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762310 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762372 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-run\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762553 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-run\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762634 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.762699 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.766389 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.766525 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.767591 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.767636 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.767631 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-dev\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.767637 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.767680 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-sys\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.767821 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b5501af-92e0-4c4a-badd-05089ae1221b-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.770137 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.771590 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.783609 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5501af-92e0-4c4a-badd-05089ae1221b-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.790160 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh57g\" (UniqueName: \"kubernetes.io/projected/7b5501af-92e0-4c4a-badd-05089ae1221b-kube-api-access-lh57g\") pod \"cinder-volume-nfs-0\" (UID: \"7b5501af-92e0-4c4a-badd-05089ae1221b\") " pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.828192 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.866471 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwws5\" (UniqueName: \"kubernetes.io/projected/e0006437-3cf9-437c-a3bd-a12f6ef84aae-kube-api-access-bwws5\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.866779 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.866804 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.866830 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.866866 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.866957 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867052 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867098 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867114 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867129 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867147 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867205 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867222 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867248 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867250 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867285 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867318 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867442 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867177 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867594 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.867614 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.868041 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.868089 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.868132 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.868608 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e0006437-3cf9-437c-a3bd-a12f6ef84aae-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.872648 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.873360 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.875737 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.879578 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0006437-3cf9-437c-a3bd-a12f6ef84aae-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.886085 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.889163 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwws5\" (UniqueName: \"kubernetes.io/projected/e0006437-3cf9-437c-a3bd-a12f6ef84aae-kube-api-access-bwws5\") pod \"cinder-volume-nfs-2-0\" (UID: \"e0006437-3cf9-437c-a3bd-a12f6ef84aae\") " pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:12 crc kubenswrapper[4789]: I1008 14:52:12.943407 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:13 crc kubenswrapper[4789]: I1008 14:52:13.461175 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 08 14:52:13 crc kubenswrapper[4789]: I1008 14:52:13.469766 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 14:52:13 crc kubenswrapper[4789]: I1008 14:52:13.575946 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Oct 08 14:52:13 crc kubenswrapper[4789]: I1008 14:52:13.601378 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"6b3acb2d-0c53-40a4-8857-8e07cf54c530","Type":"ContainerStarted","Data":"32870f71d979c1213729eb2f771161097306c7d420cec24fab3eabe86163b50c"} Oct 08 14:52:13 crc kubenswrapper[4789]: W1008 14:52:13.645118 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b5501af_92e0_4c4a_badd_05089ae1221b.slice/crio-52c22502b129b6b9f46fb1247e69d160481fb8e85e50f8c105fc37e917e380f5 WatchSource:0}: Error finding container 52c22502b129b6b9f46fb1247e69d160481fb8e85e50f8c105fc37e917e380f5: Status 404 returned error can't find the container with id 52c22502b129b6b9f46fb1247e69d160481fb8e85e50f8c105fc37e917e380f5 Oct 08 14:52:13 crc kubenswrapper[4789]: I1008 14:52:13.700937 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Oct 08 14:52:13 crc kubenswrapper[4789]: W1008 14:52:13.714297 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0006437_3cf9_437c_a3bd_a12f6ef84aae.slice/crio-bfb4610e323d56f5611b21df4860a754e7c38b3632544818d0c46617d2543a52 WatchSource:0}: Error finding container bfb4610e323d56f5611b21df4860a754e7c38b3632544818d0c46617d2543a52: Status 404 returned error can't find the container with id bfb4610e323d56f5611b21df4860a754e7c38b3632544818d0c46617d2543a52 Oct 08 14:52:14 crc kubenswrapper[4789]: I1008 14:52:14.616485 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"e0006437-3cf9-437c-a3bd-a12f6ef84aae","Type":"ContainerStarted","Data":"84782f60b585e2531d6e41c5e826fc8022052637b893d920968a4efec62abb8a"} Oct 08 14:52:14 crc kubenswrapper[4789]: I1008 14:52:14.617076 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"e0006437-3cf9-437c-a3bd-a12f6ef84aae","Type":"ContainerStarted","Data":"bfb4610e323d56f5611b21df4860a754e7c38b3632544818d0c46617d2543a52"} Oct 08 14:52:14 crc kubenswrapper[4789]: I1008 14:52:14.619192 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"6b3acb2d-0c53-40a4-8857-8e07cf54c530","Type":"ContainerStarted","Data":"cfd51a1f90ec2122ea2af54e9743ebe00b50db36180eb249d573e1f0a65b2bec"} Oct 08 14:52:14 crc kubenswrapper[4789]: I1008 14:52:14.619240 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"6b3acb2d-0c53-40a4-8857-8e07cf54c530","Type":"ContainerStarted","Data":"f9703488388e9a558fb27c5031cf46e7750202e04e7ddf572379847d61fc2f26"} Oct 08 14:52:14 crc kubenswrapper[4789]: I1008 14:52:14.620764 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"7b5501af-92e0-4c4a-badd-05089ae1221b","Type":"ContainerStarted","Data":"f78370b90b795e8664b4dd0525458be7dc961ca1fcd9d622c9a5aa38680df1df"} Oct 08 14:52:14 crc kubenswrapper[4789]: I1008 14:52:14.620809 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"7b5501af-92e0-4c4a-badd-05089ae1221b","Type":"ContainerStarted","Data":"52c22502b129b6b9f46fb1247e69d160481fb8e85e50f8c105fc37e917e380f5"} Oct 08 14:52:14 crc kubenswrapper[4789]: I1008 14:52:14.652018 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.422059563 podStartE2EDuration="2.651998817s" podCreationTimestamp="2025-10-08 14:52:12 +0000 UTC" firstStartedPulling="2025-10-08 14:52:13.469519646 +0000 UTC m=+3073.376267138" lastFinishedPulling="2025-10-08 14:52:13.6994589 +0000 UTC m=+3073.606206392" observedRunningTime="2025-10-08 14:52:14.643940287 +0000 UTC m=+3074.550687779" watchObservedRunningTime="2025-10-08 14:52:14.651998817 +0000 UTC m=+3074.558746299" Oct 08 14:52:14 crc kubenswrapper[4789]: I1008 14:52:14.680099 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-0" podStartSLOduration=2.4185977100000002 podStartE2EDuration="2.680075133s" podCreationTimestamp="2025-10-08 14:52:12 +0000 UTC" firstStartedPulling="2025-10-08 14:52:13.647851332 +0000 UTC m=+3073.554598834" lastFinishedPulling="2025-10-08 14:52:13.909328755 +0000 UTC m=+3073.816076257" observedRunningTime="2025-10-08 14:52:14.667179671 +0000 UTC m=+3074.573927173" watchObservedRunningTime="2025-10-08 14:52:14.680075133 +0000 UTC m=+3074.586822635" Oct 08 14:52:15 crc kubenswrapper[4789]: I1008 14:52:15.631501 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"e0006437-3cf9-437c-a3bd-a12f6ef84aae","Type":"ContainerStarted","Data":"d845d3f194aacc627978cf7325cbe9b77e4108baf41d83c63157f3c67b6e362c"} Oct 08 14:52:15 crc kubenswrapper[4789]: I1008 14:52:15.633784 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"7b5501af-92e0-4c4a-badd-05089ae1221b","Type":"ContainerStarted","Data":"1d45c1504643eb7fc12ff792804c8e91b1293500a1270f5ccd65ec47a1ee0bcd"} Oct 08 14:52:15 crc kubenswrapper[4789]: I1008 14:52:15.663096 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-2-0" podStartSLOduration=3.456058003 podStartE2EDuration="3.66307466s" podCreationTimestamp="2025-10-08 14:52:12 +0000 UTC" firstStartedPulling="2025-10-08 14:52:13.716574126 +0000 UTC m=+3073.623321629" lastFinishedPulling="2025-10-08 14:52:13.923590794 +0000 UTC m=+3073.830338286" observedRunningTime="2025-10-08 14:52:15.659724148 +0000 UTC m=+3075.566471640" watchObservedRunningTime="2025-10-08 14:52:15.66307466 +0000 UTC m=+3075.569822162" Oct 08 14:52:17 crc kubenswrapper[4789]: I1008 14:52:17.829079 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 08 14:52:17 crc kubenswrapper[4789]: I1008 14:52:17.886858 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:17 crc kubenswrapper[4789]: I1008 14:52:17.944198 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:22 crc kubenswrapper[4789]: I1008 14:52:22.731003 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:52:22 crc kubenswrapper[4789]: E1008 14:52:22.731760 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:52:23 crc kubenswrapper[4789]: I1008 14:52:23.050069 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 08 14:52:23 crc kubenswrapper[4789]: I1008 14:52:23.116513 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-0" Oct 08 14:52:23 crc kubenswrapper[4789]: I1008 14:52:23.160560 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-2-0" Oct 08 14:52:35 crc kubenswrapper[4789]: I1008 14:52:35.730075 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:52:35 crc kubenswrapper[4789]: E1008 14:52:35.731042 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:52:49 crc kubenswrapper[4789]: I1008 14:52:49.729733 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:52:49 crc kubenswrapper[4789]: E1008 14:52:49.730427 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:53:00 crc kubenswrapper[4789]: I1008 14:53:00.736271 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:53:00 crc kubenswrapper[4789]: E1008 14:53:00.737075 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:53:13 crc kubenswrapper[4789]: I1008 14:53:13.730787 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:53:13 crc kubenswrapper[4789]: E1008 14:53:13.731595 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:53:15 crc kubenswrapper[4789]: I1008 14:53:15.064428 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:53:15 crc kubenswrapper[4789]: I1008 14:53:15.065138 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="prometheus" containerID="cri-o://4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705" gracePeriod=600 Oct 08 14:53:15 crc kubenswrapper[4789]: I1008 14:53:15.065305 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="thanos-sidecar" containerID="cri-o://0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465" gracePeriod=600 Oct 08 14:53:15 crc kubenswrapper[4789]: I1008 14:53:15.065370 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="config-reloader" containerID="cri-o://c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c" gracePeriod=600 Oct 08 14:53:15 crc kubenswrapper[4789]: I1008 14:53:15.294560 4789 generic.go:334] "Generic (PLEG): container finished" podID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerID="0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465" exitCode=0 Oct 08 14:53:15 crc kubenswrapper[4789]: I1008 14:53:15.294627 4789 generic.go:334] "Generic (PLEG): container finished" podID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerID="4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705" exitCode=0 Oct 08 14:53:15 crc kubenswrapper[4789]: I1008 14:53:15.294646 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerDied","Data":"0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465"} Oct 08 14:53:15 crc kubenswrapper[4789]: I1008 14:53:15.294670 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerDied","Data":"4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705"} Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.063417 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112205 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-secret-combined-ca-bundle\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112283 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw6j4\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-kube-api-access-zw6j4\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112317 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-config\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112375 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112395 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c9e486c-9ff9-45f3-860e-727852ae4461-config-out\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112502 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-tls-assets\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112525 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112566 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-thanos-prometheus-http-client-file\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112615 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112755 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.112839 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0c9e486c-9ff9-45f3-860e-727852ae4461-prometheus-metric-storage-rulefiles-0\") pod \"0c9e486c-9ff9-45f3-860e-727852ae4461\" (UID: \"0c9e486c-9ff9-45f3-860e-727852ae4461\") " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.113624 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c9e486c-9ff9-45f3-860e-727852ae4461-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.122364 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.122895 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.124250 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.124592 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.124618 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-config" (OuterVolumeSpecName: "config") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.126295 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.139822 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c9e486c-9ff9-45f3-860e-727852ae4461-config-out" (OuterVolumeSpecName: "config-out") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.142556 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-kube-api-access-zw6j4" (OuterVolumeSpecName: "kube-api-access-zw6j4") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "kube-api-access-zw6j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.155467 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215027 4789 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215068 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw6j4\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-kube-api-access-zw6j4\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215083 4789 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215097 4789 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215113 4789 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c9e486c-9ff9-45f3-860e-727852ae4461-config-out\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215125 4789 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c9e486c-9ff9-45f3-860e-727852ae4461-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215139 4789 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215154 4789 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215195 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") on node \"crc\" " Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.215212 4789 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0c9e486c-9ff9-45f3-860e-727852ae4461-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.216068 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config" (OuterVolumeSpecName: "web-config") pod "0c9e486c-9ff9-45f3-860e-727852ae4461" (UID: "0c9e486c-9ff9-45f3-860e-727852ae4461"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.248031 4789 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.248741 4789 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019") on node "crc" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.317165 4789 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c9e486c-9ff9-45f3-860e-727852ae4461-web-config\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.317207 4789 reconciler_common.go:293] "Volume detached for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") on node \"crc\" DevicePath \"\"" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.320354 4789 generic.go:334] "Generic (PLEG): container finished" podID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerID="c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c" exitCode=0 Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.320405 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerDied","Data":"c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c"} Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.320441 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0c9e486c-9ff9-45f3-860e-727852ae4461","Type":"ContainerDied","Data":"40eb210931b29358f97772f27f14fe98d53668bdeebb2e745fa80d2f4990d396"} Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.320463 4789 scope.go:117] "RemoveContainer" containerID="0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.320489 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.353553 4789 scope.go:117] "RemoveContainer" containerID="c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.372269 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.391867 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.394326 4789 scope.go:117] "RemoveContainer" containerID="4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.408941 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:53:16 crc kubenswrapper[4789]: E1008 14:53:16.409739 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="config-reloader" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.409849 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="config-reloader" Oct 08 14:53:16 crc kubenswrapper[4789]: E1008 14:53:16.409949 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="init-config-reloader" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.410064 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="init-config-reloader" Oct 08 14:53:16 crc kubenswrapper[4789]: E1008 14:53:16.410182 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="prometheus" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.410275 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="prometheus" Oct 08 14:53:16 crc kubenswrapper[4789]: E1008 14:53:16.410369 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="thanos-sidecar" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.410454 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="thanos-sidecar" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.410758 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="config-reloader" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.410872 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="thanos-sidecar" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.410963 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" containerName="prometheus" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.426692 4789 scope.go:117] "RemoveContainer" containerID="f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.467197 4789 scope.go:117] "RemoveContainer" containerID="0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465" Oct 08 14:53:16 crc kubenswrapper[4789]: E1008 14:53:16.469052 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465\": container with ID starting with 0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465 not found: ID does not exist" containerID="0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.469162 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465"} err="failed to get container status \"0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465\": rpc error: code = NotFound desc = could not find container \"0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465\": container with ID starting with 0470726ca6bc43af747fc4307ffe5483c38de73b60167218cc78e109d35b7465 not found: ID does not exist" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.469199 4789 scope.go:117] "RemoveContainer" containerID="c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.469597 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.469758 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: E1008 14:53:16.469865 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c\": container with ID starting with c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c not found: ID does not exist" containerID="c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.469924 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c"} err="failed to get container status \"c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c\": rpc error: code = NotFound desc = could not find container \"c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c\": container with ID starting with c871aeff0c7d28f30379bb1d11d0068e0c99aad142ce842ad974811abcc5b00c not found: ID does not exist" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.469945 4789 scope.go:117] "RemoveContainer" containerID="4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705" Oct 08 14:53:16 crc kubenswrapper[4789]: E1008 14:53:16.472274 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705\": container with ID starting with 4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705 not found: ID does not exist" containerID="4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.482146 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705"} err="failed to get container status \"4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705\": rpc error: code = NotFound desc = could not find container \"4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705\": container with ID starting with 4f0e6f29e69e26ff80a5dd8ce1ee65f7a63c5e4fa2659475c68910b8f7f73705 not found: ID does not exist" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.482210 4789 scope.go:117] "RemoveContainer" containerID="f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.477706 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.477747 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.477833 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.478626 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-khr4j" Oct 08 14:53:16 crc kubenswrapper[4789]: E1008 14:53:16.486143 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098\": container with ID starting with f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098 not found: ID does not exist" containerID="f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.486216 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098"} err="failed to get container status \"f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098\": rpc error: code = NotFound desc = could not find container \"f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098\": container with ID starting with f850c0d6153e059e093d6d2c2b5f0fc961b63064150d5ea5ca86d5b9e9eb2098 not found: ID does not exist" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.503350 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.515578 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557222 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557284 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/441b4491-1181-4811-af7f-07e8e672352f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557354 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69v8n\" (UniqueName: \"kubernetes.io/projected/441b4491-1181-4811-af7f-07e8e672352f-kube-api-access-69v8n\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557436 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557463 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557491 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/441b4491-1181-4811-af7f-07e8e672352f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557540 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557584 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-config\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557630 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/441b4491-1181-4811-af7f-07e8e672352f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557660 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.557686 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660014 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660066 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-config\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660131 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/441b4491-1181-4811-af7f-07e8e672352f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660165 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660186 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660251 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660275 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/441b4491-1181-4811-af7f-07e8e672352f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660323 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69v8n\" (UniqueName: \"kubernetes.io/projected/441b4491-1181-4811-af7f-07e8e672352f-kube-api-access-69v8n\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660385 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660408 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.660429 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/441b4491-1181-4811-af7f-07e8e672352f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.662478 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/441b4491-1181-4811-af7f-07e8e672352f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.665102 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-config\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.665160 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.665477 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.666408 4789 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.666505 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f75c8fbff2adebf98b3c9ad4427807236950b84a4f856fba0192ffb272b3d6c1/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.667745 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.668583 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/441b4491-1181-4811-af7f-07e8e672352f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.669005 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.669017 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/441b4491-1181-4811-af7f-07e8e672352f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.680914 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/441b4491-1181-4811-af7f-07e8e672352f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.698477 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69v8n\" (UniqueName: \"kubernetes.io/projected/441b4491-1181-4811-af7f-07e8e672352f-kube-api-access-69v8n\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.721523 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-18b630d7-88dc-4a67-ba71-c2aa74e88019\") pod \"prometheus-metric-storage-0\" (UID: \"441b4491-1181-4811-af7f-07e8e672352f\") " pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.747148 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9e486c-9ff9-45f3-860e-727852ae4461" path="/var/lib/kubelet/pods/0c9e486c-9ff9-45f3-860e-727852ae4461/volumes" Oct 08 14:53:16 crc kubenswrapper[4789]: I1008 14:53:16.829928 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:17 crc kubenswrapper[4789]: I1008 14:53:17.298153 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 14:53:17 crc kubenswrapper[4789]: I1008 14:53:17.331588 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"441b4491-1181-4811-af7f-07e8e672352f","Type":"ContainerStarted","Data":"5b870af1def5dcb7de1b7f5a75c24706f52f1cc36cdbcf4cd6fb8c2d392772d8"} Oct 08 14:53:21 crc kubenswrapper[4789]: I1008 14:53:21.370785 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"441b4491-1181-4811-af7f-07e8e672352f","Type":"ContainerStarted","Data":"a869db03b0c537f5c5b93e029816b81798f44e5f1b721c8a4c80a8c247ff8b11"} Oct 08 14:53:24 crc kubenswrapper[4789]: I1008 14:53:24.730652 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:53:24 crc kubenswrapper[4789]: E1008 14:53:24.731615 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:53:29 crc kubenswrapper[4789]: I1008 14:53:29.450120 4789 generic.go:334] "Generic (PLEG): container finished" podID="441b4491-1181-4811-af7f-07e8e672352f" containerID="a869db03b0c537f5c5b93e029816b81798f44e5f1b721c8a4c80a8c247ff8b11" exitCode=0 Oct 08 14:53:29 crc kubenswrapper[4789]: I1008 14:53:29.450222 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"441b4491-1181-4811-af7f-07e8e672352f","Type":"ContainerDied","Data":"a869db03b0c537f5c5b93e029816b81798f44e5f1b721c8a4c80a8c247ff8b11"} Oct 08 14:53:30 crc kubenswrapper[4789]: I1008 14:53:30.471723 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"441b4491-1181-4811-af7f-07e8e672352f","Type":"ContainerStarted","Data":"856292e550c6958da70337bb1a8fe9a91504aae8d98be9e051395c5bab31cbda"} Oct 08 14:53:33 crc kubenswrapper[4789]: I1008 14:53:33.513486 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"441b4491-1181-4811-af7f-07e8e672352f","Type":"ContainerStarted","Data":"a9ce5f4909db3e5e8172e198c103027b0d01aeb3b227ec2cec5f5430b3487a6a"} Oct 08 14:53:33 crc kubenswrapper[4789]: I1008 14:53:33.514234 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"441b4491-1181-4811-af7f-07e8e672352f","Type":"ContainerStarted","Data":"dca3b6f8f8cd091e6cbb601493166ae27b6bd301c0b9429c6cbe587c4b5248f2"} Oct 08 14:53:33 crc kubenswrapper[4789]: I1008 14:53:33.550893 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.550876858 podStartE2EDuration="17.550876858s" podCreationTimestamp="2025-10-08 14:53:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 14:53:33.544287948 +0000 UTC m=+3153.451035450" watchObservedRunningTime="2025-10-08 14:53:33.550876858 +0000 UTC m=+3153.457624350" Oct 08 14:53:36 crc kubenswrapper[4789]: I1008 14:53:36.730462 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:53:36 crc kubenswrapper[4789]: E1008 14:53:36.731238 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:53:36 crc kubenswrapper[4789]: I1008 14:53:36.830587 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:46 crc kubenswrapper[4789]: I1008 14:53:46.830959 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:46 crc kubenswrapper[4789]: I1008 14:53:46.840501 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:47 crc kubenswrapper[4789]: I1008 14:53:47.655422 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 08 14:53:47 crc kubenswrapper[4789]: I1008 14:53:47.730757 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:53:47 crc kubenswrapper[4789]: E1008 14:53:47.731043 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:54:01 crc kubenswrapper[4789]: I1008 14:54:01.730340 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:54:01 crc kubenswrapper[4789]: E1008 14:54:01.731973 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:54:13 crc kubenswrapper[4789]: I1008 14:54:13.841111 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 14:54:13 crc kubenswrapper[4789]: I1008 14:54:13.843376 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 14:54:13 crc kubenswrapper[4789]: I1008 14:54:13.846777 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 08 14:54:13 crc kubenswrapper[4789]: I1008 14:54:13.846895 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-thkkz" Oct 08 14:54:13 crc kubenswrapper[4789]: I1008 14:54:13.847211 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 08 14:54:13 crc kubenswrapper[4789]: I1008 14:54:13.847517 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 08 14:54:13 crc kubenswrapper[4789]: I1008 14:54:13.858255 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.033280 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.033383 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.033459 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.033597 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-config-data\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.033816 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.033912 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.033931 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.034038 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpd2s\" (UniqueName: \"kubernetes.io/projected/5622aa64-be18-4420-8c84-b76fe03c1346-kube-api-access-zpd2s\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.034094 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.135682 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-config-data\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.135749 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.135788 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.135818 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.135899 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpd2s\" (UniqueName: \"kubernetes.io/projected/5622aa64-be18-4420-8c84-b76fe03c1346-kube-api-access-zpd2s\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.135928 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.136059 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.136102 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.136144 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.136412 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.136540 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.136714 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.136966 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-config-data\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.137194 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.149268 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.151968 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpd2s\" (UniqueName: \"kubernetes.io/projected/5622aa64-be18-4420-8c84-b76fe03c1346-kube-api-access-zpd2s\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.153766 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.157065 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.202068 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " pod="openstack/tempest-tests-tempest" Oct 08 14:54:14 crc kubenswrapper[4789]: I1008 14:54:14.468485 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 14:54:15 crc kubenswrapper[4789]: I1008 14:54:15.584342 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 14:54:15 crc kubenswrapper[4789]: I1008 14:54:15.911291 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5622aa64-be18-4420-8c84-b76fe03c1346","Type":"ContainerStarted","Data":"f18c299b113e7a48a2c3dfe5a9dc13a9ad3198cd51144a885c971a52a8934235"} Oct 08 14:54:16 crc kubenswrapper[4789]: I1008 14:54:16.730161 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:54:16 crc kubenswrapper[4789]: E1008 14:54:16.730654 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:54:27 crc kubenswrapper[4789]: I1008 14:54:27.007730 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5622aa64-be18-4420-8c84-b76fe03c1346","Type":"ContainerStarted","Data":"84e3bd795069144ca408b979b068ed5ffe872cc3f6b473b080b59c236d61ecb4"} Oct 08 14:54:27 crc kubenswrapper[4789]: I1008 14:54:27.027206 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.889769685 podStartE2EDuration="15.027191051s" podCreationTimestamp="2025-10-08 14:54:12 +0000 UTC" firstStartedPulling="2025-10-08 14:54:15.590651892 +0000 UTC m=+3195.497399414" lastFinishedPulling="2025-10-08 14:54:25.728073258 +0000 UTC m=+3205.634820780" observedRunningTime="2025-10-08 14:54:27.023730776 +0000 UTC m=+3206.930478288" watchObservedRunningTime="2025-10-08 14:54:27.027191051 +0000 UTC m=+3206.933938543" Oct 08 14:54:28 crc kubenswrapper[4789]: I1008 14:54:28.732174 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:54:28 crc kubenswrapper[4789]: E1008 14:54:28.732694 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:54:43 crc kubenswrapper[4789]: I1008 14:54:43.730267 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:54:43 crc kubenswrapper[4789]: E1008 14:54:43.731072 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:54:56 crc kubenswrapper[4789]: I1008 14:54:56.730799 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:54:56 crc kubenswrapper[4789]: E1008 14:54:56.731884 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:55:07 crc kubenswrapper[4789]: I1008 14:55:07.730931 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:55:07 crc kubenswrapper[4789]: E1008 14:55:07.732184 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:55:18 crc kubenswrapper[4789]: I1008 14:55:18.730589 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:55:18 crc kubenswrapper[4789]: E1008 14:55:18.731349 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 14:55:32 crc kubenswrapper[4789]: I1008 14:55:32.730564 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:55:33 crc kubenswrapper[4789]: I1008 14:55:33.603716 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"03559d7c0bfab1b8eaf3db8cb07705d99d1ef3a70471e25a70558c926b95793c"} Oct 08 14:56:25 crc kubenswrapper[4789]: I1008 14:56:25.875914 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hx5f6"] Oct 08 14:56:25 crc kubenswrapper[4789]: I1008 14:56:25.878902 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:25 crc kubenswrapper[4789]: I1008 14:56:25.889208 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hx5f6"] Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.007829 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-utilities\") pod \"community-operators-hx5f6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.007887 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mdhh\" (UniqueName: \"kubernetes.io/projected/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-kube-api-access-6mdhh\") pod \"community-operators-hx5f6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.007912 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-catalog-content\") pod \"community-operators-hx5f6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.110234 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-utilities\") pod \"community-operators-hx5f6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.110307 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mdhh\" (UniqueName: \"kubernetes.io/projected/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-kube-api-access-6mdhh\") pod \"community-operators-hx5f6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.110334 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-catalog-content\") pod \"community-operators-hx5f6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.110683 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-utilities\") pod \"community-operators-hx5f6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.110818 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-catalog-content\") pod \"community-operators-hx5f6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.130247 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mdhh\" (UniqueName: \"kubernetes.io/projected/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-kube-api-access-6mdhh\") pod \"community-operators-hx5f6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.209907 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:26 crc kubenswrapper[4789]: I1008 14:56:26.756909 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hx5f6"] Oct 08 14:56:27 crc kubenswrapper[4789]: I1008 14:56:27.124546 4789 generic.go:334] "Generic (PLEG): container finished" podID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerID="a929595bebde030e07ee01ffaafc1d0f14a6ca599c11425f696f7618ffef57bc" exitCode=0 Oct 08 14:56:27 crc kubenswrapper[4789]: I1008 14:56:27.124594 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx5f6" event={"ID":"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6","Type":"ContainerDied","Data":"a929595bebde030e07ee01ffaafc1d0f14a6ca599c11425f696f7618ffef57bc"} Oct 08 14:56:27 crc kubenswrapper[4789]: I1008 14:56:27.124915 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx5f6" event={"ID":"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6","Type":"ContainerStarted","Data":"0db6ad268cd6acfdfcca6afac1f83ffa5c36e2603c350f7cc966e5ce9968addd"} Oct 08 14:56:29 crc kubenswrapper[4789]: I1008 14:56:29.200540 4789 generic.go:334] "Generic (PLEG): container finished" podID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerID="a57e8a500c230d6969337f6f8a9ea518499c5f1bb0f4238009dd18a41b64e144" exitCode=0 Oct 08 14:56:29 crc kubenswrapper[4789]: I1008 14:56:29.201339 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx5f6" event={"ID":"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6","Type":"ContainerDied","Data":"a57e8a500c230d6969337f6f8a9ea518499c5f1bb0f4238009dd18a41b64e144"} Oct 08 14:56:29 crc kubenswrapper[4789]: I1008 14:56:29.878756 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mgc8q"] Oct 08 14:56:29 crc kubenswrapper[4789]: I1008 14:56:29.881513 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:29 crc kubenswrapper[4789]: I1008 14:56:29.893411 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgc8q"] Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.002194 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-catalog-content\") pod \"redhat-marketplace-mgc8q\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.002369 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-utilities\") pod \"redhat-marketplace-mgc8q\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.002411 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9s2l\" (UniqueName: \"kubernetes.io/projected/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-kube-api-access-z9s2l\") pod \"redhat-marketplace-mgc8q\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.105312 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-utilities\") pod \"redhat-marketplace-mgc8q\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.105395 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9s2l\" (UniqueName: \"kubernetes.io/projected/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-kube-api-access-z9s2l\") pod \"redhat-marketplace-mgc8q\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.105485 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-catalog-content\") pod \"redhat-marketplace-mgc8q\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.106213 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-catalog-content\") pod \"redhat-marketplace-mgc8q\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.106541 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-utilities\") pod \"redhat-marketplace-mgc8q\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.130325 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9s2l\" (UniqueName: \"kubernetes.io/projected/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-kube-api-access-z9s2l\") pod \"redhat-marketplace-mgc8q\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.199788 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:30 crc kubenswrapper[4789]: I1008 14:56:30.747574 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgc8q"] Oct 08 14:56:31 crc kubenswrapper[4789]: I1008 14:56:31.224400 4789 generic.go:334] "Generic (PLEG): container finished" podID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerID="ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519" exitCode=0 Oct 08 14:56:31 crc kubenswrapper[4789]: I1008 14:56:31.224443 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgc8q" event={"ID":"632fc91b-0def-4f9a-9a8a-9b62566c4e0c","Type":"ContainerDied","Data":"ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519"} Oct 08 14:56:31 crc kubenswrapper[4789]: I1008 14:56:31.224764 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgc8q" event={"ID":"632fc91b-0def-4f9a-9a8a-9b62566c4e0c","Type":"ContainerStarted","Data":"2dd50e36d1b25f203dd74f514200986c58d75b9ee670dc8129549c27673ac3fa"} Oct 08 14:56:31 crc kubenswrapper[4789]: I1008 14:56:31.229366 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx5f6" event={"ID":"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6","Type":"ContainerStarted","Data":"dec10a416c9e1898d2aa7a401ccff0d83174008ee344a43e2c08c06d014dc46c"} Oct 08 14:56:31 crc kubenswrapper[4789]: I1008 14:56:31.281306 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hx5f6" podStartSLOduration=2.631845714 podStartE2EDuration="6.281284476s" podCreationTimestamp="2025-10-08 14:56:25 +0000 UTC" firstStartedPulling="2025-10-08 14:56:27.127462903 +0000 UTC m=+3327.034210395" lastFinishedPulling="2025-10-08 14:56:30.776901665 +0000 UTC m=+3330.683649157" observedRunningTime="2025-10-08 14:56:31.279375404 +0000 UTC m=+3331.186122896" watchObservedRunningTime="2025-10-08 14:56:31.281284476 +0000 UTC m=+3331.188031978" Oct 08 14:56:33 crc kubenswrapper[4789]: I1008 14:56:33.281761 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgc8q" event={"ID":"632fc91b-0def-4f9a-9a8a-9b62566c4e0c","Type":"ContainerStarted","Data":"1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed"} Oct 08 14:56:35 crc kubenswrapper[4789]: I1008 14:56:35.299816 4789 generic.go:334] "Generic (PLEG): container finished" podID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerID="1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed" exitCode=0 Oct 08 14:56:35 crc kubenswrapper[4789]: I1008 14:56:35.299905 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgc8q" event={"ID":"632fc91b-0def-4f9a-9a8a-9b62566c4e0c","Type":"ContainerDied","Data":"1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed"} Oct 08 14:56:36 crc kubenswrapper[4789]: I1008 14:56:36.210426 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:36 crc kubenswrapper[4789]: I1008 14:56:36.210681 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:36 crc kubenswrapper[4789]: I1008 14:56:36.257060 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:36 crc kubenswrapper[4789]: I1008 14:56:36.333700 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgc8q" event={"ID":"632fc91b-0def-4f9a-9a8a-9b62566c4e0c","Type":"ContainerStarted","Data":"71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85"} Oct 08 14:56:36 crc kubenswrapper[4789]: I1008 14:56:36.356528 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mgc8q" podStartSLOduration=2.8051890779999997 podStartE2EDuration="7.356509686s" podCreationTimestamp="2025-10-08 14:56:29 +0000 UTC" firstStartedPulling="2025-10-08 14:56:31.226333557 +0000 UTC m=+3331.133081049" lastFinishedPulling="2025-10-08 14:56:35.777654165 +0000 UTC m=+3335.684401657" observedRunningTime="2025-10-08 14:56:36.352007113 +0000 UTC m=+3336.258754605" watchObservedRunningTime="2025-10-08 14:56:36.356509686 +0000 UTC m=+3336.263257178" Oct 08 14:56:36 crc kubenswrapper[4789]: I1008 14:56:36.388421 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:38 crc kubenswrapper[4789]: I1008 14:56:38.867466 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hx5f6"] Oct 08 14:56:38 crc kubenswrapper[4789]: I1008 14:56:38.868152 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hx5f6" podUID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerName="registry-server" containerID="cri-o://dec10a416c9e1898d2aa7a401ccff0d83174008ee344a43e2c08c06d014dc46c" gracePeriod=2 Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.361492 4789 generic.go:334] "Generic (PLEG): container finished" podID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerID="dec10a416c9e1898d2aa7a401ccff0d83174008ee344a43e2c08c06d014dc46c" exitCode=0 Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.361546 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx5f6" event={"ID":"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6","Type":"ContainerDied","Data":"dec10a416c9e1898d2aa7a401ccff0d83174008ee344a43e2c08c06d014dc46c"} Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.361870 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx5f6" event={"ID":"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6","Type":"ContainerDied","Data":"0db6ad268cd6acfdfcca6afac1f83ffa5c36e2603c350f7cc966e5ce9968addd"} Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.361887 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0db6ad268cd6acfdfcca6afac1f83ffa5c36e2603c350f7cc966e5ce9968addd" Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.444949 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.599864 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-catalog-content\") pod \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.599917 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-utilities\") pod \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.600115 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mdhh\" (UniqueName: \"kubernetes.io/projected/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-kube-api-access-6mdhh\") pod \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\" (UID: \"d2cc87a4-b5b8-4fe1-896e-bbe520224ea6\") " Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.600612 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-utilities" (OuterVolumeSpecName: "utilities") pod "d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" (UID: "d2cc87a4-b5b8-4fe1-896e-bbe520224ea6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.608344 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-kube-api-access-6mdhh" (OuterVolumeSpecName: "kube-api-access-6mdhh") pod "d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" (UID: "d2cc87a4-b5b8-4fe1-896e-bbe520224ea6"). InnerVolumeSpecName "kube-api-access-6mdhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.652315 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" (UID: "d2cc87a4-b5b8-4fe1-896e-bbe520224ea6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.702618 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.702650 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:56:39 crc kubenswrapper[4789]: I1008 14:56:39.702664 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mdhh\" (UniqueName: \"kubernetes.io/projected/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6-kube-api-access-6mdhh\") on node \"crc\" DevicePath \"\"" Oct 08 14:56:40 crc kubenswrapper[4789]: I1008 14:56:40.200799 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:40 crc kubenswrapper[4789]: I1008 14:56:40.201129 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:40 crc kubenswrapper[4789]: I1008 14:56:40.249422 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:40 crc kubenswrapper[4789]: I1008 14:56:40.371232 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx5f6" Oct 08 14:56:40 crc kubenswrapper[4789]: I1008 14:56:40.404344 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hx5f6"] Oct 08 14:56:40 crc kubenswrapper[4789]: I1008 14:56:40.412463 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hx5f6"] Oct 08 14:56:40 crc kubenswrapper[4789]: I1008 14:56:40.428645 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:40 crc kubenswrapper[4789]: I1008 14:56:40.743935 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" path="/var/lib/kubelet/pods/d2cc87a4-b5b8-4fe1-896e-bbe520224ea6/volumes" Oct 08 14:56:42 crc kubenswrapper[4789]: I1008 14:56:42.069055 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgc8q"] Oct 08 14:56:42 crc kubenswrapper[4789]: I1008 14:56:42.386616 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mgc8q" podUID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerName="registry-server" containerID="cri-o://71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85" gracePeriod=2 Oct 08 14:56:42 crc kubenswrapper[4789]: I1008 14:56:42.901162 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:42 crc kubenswrapper[4789]: I1008 14:56:42.977345 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-utilities\") pod \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " Oct 08 14:56:42 crc kubenswrapper[4789]: I1008 14:56:42.977625 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-catalog-content\") pod \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " Oct 08 14:56:42 crc kubenswrapper[4789]: I1008 14:56:42.977658 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9s2l\" (UniqueName: \"kubernetes.io/projected/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-kube-api-access-z9s2l\") pod \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\" (UID: \"632fc91b-0def-4f9a-9a8a-9b62566c4e0c\") " Oct 08 14:56:42 crc kubenswrapper[4789]: I1008 14:56:42.978293 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-utilities" (OuterVolumeSpecName: "utilities") pod "632fc91b-0def-4f9a-9a8a-9b62566c4e0c" (UID: "632fc91b-0def-4f9a-9a8a-9b62566c4e0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:56:42 crc kubenswrapper[4789]: I1008 14:56:42.988809 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-kube-api-access-z9s2l" (OuterVolumeSpecName: "kube-api-access-z9s2l") pod "632fc91b-0def-4f9a-9a8a-9b62566c4e0c" (UID: "632fc91b-0def-4f9a-9a8a-9b62566c4e0c"). InnerVolumeSpecName "kube-api-access-z9s2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 14:56:42 crc kubenswrapper[4789]: I1008 14:56:42.995802 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "632fc91b-0def-4f9a-9a8a-9b62566c4e0c" (UID: "632fc91b-0def-4f9a-9a8a-9b62566c4e0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.080157 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.080189 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.080203 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9s2l\" (UniqueName: \"kubernetes.io/projected/632fc91b-0def-4f9a-9a8a-9b62566c4e0c-kube-api-access-z9s2l\") on node \"crc\" DevicePath \"\"" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.397379 4789 generic.go:334] "Generic (PLEG): container finished" podID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerID="71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85" exitCode=0 Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.397512 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgc8q" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.397538 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgc8q" event={"ID":"632fc91b-0def-4f9a-9a8a-9b62566c4e0c","Type":"ContainerDied","Data":"71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85"} Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.398678 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgc8q" event={"ID":"632fc91b-0def-4f9a-9a8a-9b62566c4e0c","Type":"ContainerDied","Data":"2dd50e36d1b25f203dd74f514200986c58d75b9ee670dc8129549c27673ac3fa"} Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.398717 4789 scope.go:117] "RemoveContainer" containerID="71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.429701 4789 scope.go:117] "RemoveContainer" containerID="1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.438952 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgc8q"] Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.460351 4789 scope.go:117] "RemoveContainer" containerID="ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.461437 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgc8q"] Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.501049 4789 scope.go:117] "RemoveContainer" containerID="71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85" Oct 08 14:56:43 crc kubenswrapper[4789]: E1008 14:56:43.501506 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85\": container with ID starting with 71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85 not found: ID does not exist" containerID="71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.501537 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85"} err="failed to get container status \"71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85\": rpc error: code = NotFound desc = could not find container \"71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85\": container with ID starting with 71b085069f57e5b8fcc0f94548bc9fbd88766e38a53bd029641d0c8c791a4c85 not found: ID does not exist" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.501561 4789 scope.go:117] "RemoveContainer" containerID="1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed" Oct 08 14:56:43 crc kubenswrapper[4789]: E1008 14:56:43.505451 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed\": container with ID starting with 1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed not found: ID does not exist" containerID="1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.505478 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed"} err="failed to get container status \"1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed\": rpc error: code = NotFound desc = could not find container \"1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed\": container with ID starting with 1777d0c091e1ca6ee63f882bc803274db4d611cd37b81485bdbe15961939dbed not found: ID does not exist" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.505499 4789 scope.go:117] "RemoveContainer" containerID="ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519" Oct 08 14:56:43 crc kubenswrapper[4789]: E1008 14:56:43.505918 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519\": container with ID starting with ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519 not found: ID does not exist" containerID="ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519" Oct 08 14:56:43 crc kubenswrapper[4789]: I1008 14:56:43.505939 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519"} err="failed to get container status \"ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519\": rpc error: code = NotFound desc = could not find container \"ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519\": container with ID starting with ed8c231afe91bc7bbc4195e53cecc0b10fdf52dcc5acd4d969b040be57830519 not found: ID does not exist" Oct 08 14:56:44 crc kubenswrapper[4789]: I1008 14:56:44.740510 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" path="/var/lib/kubelet/pods/632fc91b-0def-4f9a-9a8a-9b62566c4e0c/volumes" Oct 08 14:57:56 crc kubenswrapper[4789]: I1008 14:57:56.432427 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:57:56 crc kubenswrapper[4789]: I1008 14:57:56.432891 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:58:26 crc kubenswrapper[4789]: I1008 14:58:26.432603 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:58:26 crc kubenswrapper[4789]: I1008 14:58:26.433125 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:58:56 crc kubenswrapper[4789]: I1008 14:58:56.433157 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 14:58:56 crc kubenswrapper[4789]: I1008 14:58:56.434866 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 14:58:56 crc kubenswrapper[4789]: I1008 14:58:56.435004 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 14:58:56 crc kubenswrapper[4789]: I1008 14:58:56.435634 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"03559d7c0bfab1b8eaf3db8cb07705d99d1ef3a70471e25a70558c926b95793c"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 14:58:56 crc kubenswrapper[4789]: I1008 14:58:56.435952 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://03559d7c0bfab1b8eaf3db8cb07705d99d1ef3a70471e25a70558c926b95793c" gracePeriod=600 Oct 08 14:58:56 crc kubenswrapper[4789]: I1008 14:58:56.596541 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="03559d7c0bfab1b8eaf3db8cb07705d99d1ef3a70471e25a70558c926b95793c" exitCode=0 Oct 08 14:58:56 crc kubenswrapper[4789]: I1008 14:58:56.596614 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"03559d7c0bfab1b8eaf3db8cb07705d99d1ef3a70471e25a70558c926b95793c"} Oct 08 14:58:56 crc kubenswrapper[4789]: I1008 14:58:56.597162 4789 scope.go:117] "RemoveContainer" containerID="6acf5f309ed88299e461065a1ec974ad9481bdd5314a17295917ce822a1f2170" Oct 08 14:58:57 crc kubenswrapper[4789]: I1008 14:58:57.612365 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea"} Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.191133 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2"] Oct 08 15:00:00 crc kubenswrapper[4789]: E1008 15:00:00.192736 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerName="registry-server" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.192755 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerName="registry-server" Oct 08 15:00:00 crc kubenswrapper[4789]: E1008 15:00:00.192784 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerName="registry-server" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.192792 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerName="registry-server" Oct 08 15:00:00 crc kubenswrapper[4789]: E1008 15:00:00.192808 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerName="extract-content" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.192821 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerName="extract-content" Oct 08 15:00:00 crc kubenswrapper[4789]: E1008 15:00:00.192840 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerName="extract-utilities" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.192849 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerName="extract-utilities" Oct 08 15:00:00 crc kubenswrapper[4789]: E1008 15:00:00.192881 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerName="extract-content" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.192893 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerName="extract-content" Oct 08 15:00:00 crc kubenswrapper[4789]: E1008 15:00:00.192922 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerName="extract-utilities" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.193033 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerName="extract-utilities" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.194681 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="632fc91b-0def-4f9a-9a8a-9b62566c4e0c" containerName="registry-server" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.194856 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2cc87a4-b5b8-4fe1-896e-bbe520224ea6" containerName="registry-server" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.196880 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.200514 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.206674 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.227432 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2"] Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.241483 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d9e7f6-2d4e-4512-91ab-0430a401721d-secret-volume\") pod \"collect-profiles-29332260-rcsz2\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.241553 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zstjc\" (UniqueName: \"kubernetes.io/projected/03d9e7f6-2d4e-4512-91ab-0430a401721d-kube-api-access-zstjc\") pod \"collect-profiles-29332260-rcsz2\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.241767 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d9e7f6-2d4e-4512-91ab-0430a401721d-config-volume\") pod \"collect-profiles-29332260-rcsz2\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.344083 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d9e7f6-2d4e-4512-91ab-0430a401721d-config-volume\") pod \"collect-profiles-29332260-rcsz2\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.344222 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d9e7f6-2d4e-4512-91ab-0430a401721d-secret-volume\") pod \"collect-profiles-29332260-rcsz2\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.344264 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zstjc\" (UniqueName: \"kubernetes.io/projected/03d9e7f6-2d4e-4512-91ab-0430a401721d-kube-api-access-zstjc\") pod \"collect-profiles-29332260-rcsz2\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.345806 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d9e7f6-2d4e-4512-91ab-0430a401721d-config-volume\") pod \"collect-profiles-29332260-rcsz2\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.350817 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d9e7f6-2d4e-4512-91ab-0430a401721d-secret-volume\") pod \"collect-profiles-29332260-rcsz2\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.360959 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zstjc\" (UniqueName: \"kubernetes.io/projected/03d9e7f6-2d4e-4512-91ab-0430a401721d-kube-api-access-zstjc\") pod \"collect-profiles-29332260-rcsz2\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.528489 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:00 crc kubenswrapper[4789]: I1008 15:00:00.975588 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2"] Oct 08 15:00:00 crc kubenswrapper[4789]: W1008 15:00:00.981827 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03d9e7f6_2d4e_4512_91ab_0430a401721d.slice/crio-1a9577d3d8e30f9385ec40eab1014c0c09c4df29709947d795f90241bb022728 WatchSource:0}: Error finding container 1a9577d3d8e30f9385ec40eab1014c0c09c4df29709947d795f90241bb022728: Status 404 returned error can't find the container with id 1a9577d3d8e30f9385ec40eab1014c0c09c4df29709947d795f90241bb022728 Oct 08 15:00:01 crc kubenswrapper[4789]: I1008 15:00:01.240747 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" event={"ID":"03d9e7f6-2d4e-4512-91ab-0430a401721d","Type":"ContainerStarted","Data":"ddbc84812cf1e0be05faf746d2bab1c1356354e5d74adf6d914da9602e3c6a43"} Oct 08 15:00:01 crc kubenswrapper[4789]: I1008 15:00:01.240822 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" event={"ID":"03d9e7f6-2d4e-4512-91ab-0430a401721d","Type":"ContainerStarted","Data":"1a9577d3d8e30f9385ec40eab1014c0c09c4df29709947d795f90241bb022728"} Oct 08 15:00:01 crc kubenswrapper[4789]: I1008 15:00:01.269817 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" podStartSLOduration=1.2697968259999999 podStartE2EDuration="1.269796826s" podCreationTimestamp="2025-10-08 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:00:01.259939877 +0000 UTC m=+3541.166687379" watchObservedRunningTime="2025-10-08 15:00:01.269796826 +0000 UTC m=+3541.176544318" Oct 08 15:00:02 crc kubenswrapper[4789]: I1008 15:00:02.256432 4789 generic.go:334] "Generic (PLEG): container finished" podID="03d9e7f6-2d4e-4512-91ab-0430a401721d" containerID="ddbc84812cf1e0be05faf746d2bab1c1356354e5d74adf6d914da9602e3c6a43" exitCode=0 Oct 08 15:00:02 crc kubenswrapper[4789]: I1008 15:00:02.256592 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" event={"ID":"03d9e7f6-2d4e-4512-91ab-0430a401721d","Type":"ContainerDied","Data":"ddbc84812cf1e0be05faf746d2bab1c1356354e5d74adf6d914da9602e3c6a43"} Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.622600 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.727871 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d9e7f6-2d4e-4512-91ab-0430a401721d-config-volume\") pod \"03d9e7f6-2d4e-4512-91ab-0430a401721d\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.728125 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zstjc\" (UniqueName: \"kubernetes.io/projected/03d9e7f6-2d4e-4512-91ab-0430a401721d-kube-api-access-zstjc\") pod \"03d9e7f6-2d4e-4512-91ab-0430a401721d\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.728249 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d9e7f6-2d4e-4512-91ab-0430a401721d-secret-volume\") pod \"03d9e7f6-2d4e-4512-91ab-0430a401721d\" (UID: \"03d9e7f6-2d4e-4512-91ab-0430a401721d\") " Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.729072 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d9e7f6-2d4e-4512-91ab-0430a401721d-config-volume" (OuterVolumeSpecName: "config-volume") pod "03d9e7f6-2d4e-4512-91ab-0430a401721d" (UID: "03d9e7f6-2d4e-4512-91ab-0430a401721d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.734980 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d9e7f6-2d4e-4512-91ab-0430a401721d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "03d9e7f6-2d4e-4512-91ab-0430a401721d" (UID: "03d9e7f6-2d4e-4512-91ab-0430a401721d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.735387 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d9e7f6-2d4e-4512-91ab-0430a401721d-kube-api-access-zstjc" (OuterVolumeSpecName: "kube-api-access-zstjc") pod "03d9e7f6-2d4e-4512-91ab-0430a401721d" (UID: "03d9e7f6-2d4e-4512-91ab-0430a401721d"). InnerVolumeSpecName "kube-api-access-zstjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.831227 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/03d9e7f6-2d4e-4512-91ab-0430a401721d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.831489 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/03d9e7f6-2d4e-4512-91ab-0430a401721d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:00:03 crc kubenswrapper[4789]: I1008 15:00:03.831567 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zstjc\" (UniqueName: \"kubernetes.io/projected/03d9e7f6-2d4e-4512-91ab-0430a401721d-kube-api-access-zstjc\") on node \"crc\" DevicePath \"\"" Oct 08 15:00:04 crc kubenswrapper[4789]: I1008 15:00:04.281909 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" event={"ID":"03d9e7f6-2d4e-4512-91ab-0430a401721d","Type":"ContainerDied","Data":"1a9577d3d8e30f9385ec40eab1014c0c09c4df29709947d795f90241bb022728"} Oct 08 15:00:04 crc kubenswrapper[4789]: I1008 15:00:04.282303 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a9577d3d8e30f9385ec40eab1014c0c09c4df29709947d795f90241bb022728" Oct 08 15:00:04 crc kubenswrapper[4789]: I1008 15:00:04.281970 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2" Oct 08 15:00:04 crc kubenswrapper[4789]: I1008 15:00:04.332837 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv"] Oct 08 15:00:04 crc kubenswrapper[4789]: I1008 15:00:04.340754 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332215-qjnsv"] Oct 08 15:00:04 crc kubenswrapper[4789]: I1008 15:00:04.747463 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee6a88e2-3621-4c9f-8902-e13dd3799854" path="/var/lib/kubelet/pods/ee6a88e2-3621-4c9f-8902-e13dd3799854/volumes" Oct 08 15:00:24 crc kubenswrapper[4789]: I1008 15:00:24.005651 4789 scope.go:117] "RemoveContainer" containerID="6a1b81d36056cbf14c53b74c5966aa4be8e99e33ade58c21a37ce6051b78cb71" Oct 08 15:00:54 crc kubenswrapper[4789]: I1008 15:00:54.786919 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h47hs"] Oct 08 15:00:54 crc kubenswrapper[4789]: E1008 15:00:54.788619 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d9e7f6-2d4e-4512-91ab-0430a401721d" containerName="collect-profiles" Oct 08 15:00:54 crc kubenswrapper[4789]: I1008 15:00:54.788642 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d9e7f6-2d4e-4512-91ab-0430a401721d" containerName="collect-profiles" Oct 08 15:00:54 crc kubenswrapper[4789]: I1008 15:00:54.788977 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="03d9e7f6-2d4e-4512-91ab-0430a401721d" containerName="collect-profiles" Oct 08 15:00:54 crc kubenswrapper[4789]: I1008 15:00:54.791498 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:54 crc kubenswrapper[4789]: I1008 15:00:54.804813 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h47hs"] Oct 08 15:00:54 crc kubenswrapper[4789]: I1008 15:00:54.950134 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-catalog-content\") pod \"certified-operators-h47hs\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:54 crc kubenswrapper[4789]: I1008 15:00:54.950250 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6m46\" (UniqueName: \"kubernetes.io/projected/4b5bc008-5174-4e13-a71f-be5f9bd756cf-kube-api-access-x6m46\") pod \"certified-operators-h47hs\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:54 crc kubenswrapper[4789]: I1008 15:00:54.950314 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-utilities\") pod \"certified-operators-h47hs\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:55 crc kubenswrapper[4789]: I1008 15:00:55.053943 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-catalog-content\") pod \"certified-operators-h47hs\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:55 crc kubenswrapper[4789]: I1008 15:00:55.054295 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6m46\" (UniqueName: \"kubernetes.io/projected/4b5bc008-5174-4e13-a71f-be5f9bd756cf-kube-api-access-x6m46\") pod \"certified-operators-h47hs\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:55 crc kubenswrapper[4789]: I1008 15:00:55.054459 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-utilities\") pod \"certified-operators-h47hs\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:55 crc kubenswrapper[4789]: I1008 15:00:55.055042 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-utilities\") pod \"certified-operators-h47hs\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:55 crc kubenswrapper[4789]: I1008 15:00:55.055262 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-catalog-content\") pod \"certified-operators-h47hs\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:55 crc kubenswrapper[4789]: I1008 15:00:55.079301 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6m46\" (UniqueName: \"kubernetes.io/projected/4b5bc008-5174-4e13-a71f-be5f9bd756cf-kube-api-access-x6m46\") pod \"certified-operators-h47hs\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:55 crc kubenswrapper[4789]: I1008 15:00:55.123159 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:00:55 crc kubenswrapper[4789]: I1008 15:00:55.513890 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h47hs"] Oct 08 15:00:55 crc kubenswrapper[4789]: I1008 15:00:55.752434 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47hs" event={"ID":"4b5bc008-5174-4e13-a71f-be5f9bd756cf","Type":"ContainerStarted","Data":"760f9fdf9b79fc6976ffa56d5bf266c4cea2aacfd573b1bdbe7713de3ddf7271"} Oct 08 15:00:56 crc kubenswrapper[4789]: I1008 15:00:56.433430 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:00:56 crc kubenswrapper[4789]: I1008 15:00:56.433841 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:00:56 crc kubenswrapper[4789]: I1008 15:00:56.762580 4789 generic.go:334] "Generic (PLEG): container finished" podID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerID="968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde" exitCode=0 Oct 08 15:00:56 crc kubenswrapper[4789]: I1008 15:00:56.762638 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47hs" event={"ID":"4b5bc008-5174-4e13-a71f-be5f9bd756cf","Type":"ContainerDied","Data":"968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde"} Oct 08 15:00:56 crc kubenswrapper[4789]: I1008 15:00:56.764758 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:00:58 crc kubenswrapper[4789]: I1008 15:00:58.791314 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47hs" event={"ID":"4b5bc008-5174-4e13-a71f-be5f9bd756cf","Type":"ContainerStarted","Data":"6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7"} Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.153508 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29332261-6nq76"] Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.155109 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.176533 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332261-6nq76"] Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.269879 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-config-data\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.269968 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-combined-ca-bundle\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.270019 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-fernet-keys\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.270106 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rw6v\" (UniqueName: \"kubernetes.io/projected/caa89bd3-cdc8-412a-82f0-1289cf0028c1-kube-api-access-6rw6v\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.373490 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-combined-ca-bundle\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.373601 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-fernet-keys\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.373696 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rw6v\" (UniqueName: \"kubernetes.io/projected/caa89bd3-cdc8-412a-82f0-1289cf0028c1-kube-api-access-6rw6v\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.373841 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-config-data\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.519280 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rw6v\" (UniqueName: \"kubernetes.io/projected/caa89bd3-cdc8-412a-82f0-1289cf0028c1-kube-api-access-6rw6v\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.519968 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-combined-ca-bundle\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.520644 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-config-data\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.531908 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-fernet-keys\") pod \"keystone-cron-29332261-6nq76\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.815903 4789 generic.go:334] "Generic (PLEG): container finished" podID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerID="6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7" exitCode=0 Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.815968 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47hs" event={"ID":"4b5bc008-5174-4e13-a71f-be5f9bd756cf","Type":"ContainerDied","Data":"6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7"} Oct 08 15:01:00 crc kubenswrapper[4789]: I1008 15:01:00.824002 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:01 crc kubenswrapper[4789]: I1008 15:01:01.318867 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332261-6nq76"] Oct 08 15:01:01 crc kubenswrapper[4789]: I1008 15:01:01.828215 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47hs" event={"ID":"4b5bc008-5174-4e13-a71f-be5f9bd756cf","Type":"ContainerStarted","Data":"270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f"} Oct 08 15:01:01 crc kubenswrapper[4789]: I1008 15:01:01.831196 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332261-6nq76" event={"ID":"caa89bd3-cdc8-412a-82f0-1289cf0028c1","Type":"ContainerStarted","Data":"5be03d1c82ee66b2fdad68abf2d4e2a2369013da360b5c7cb40b0330b039185e"} Oct 08 15:01:01 crc kubenswrapper[4789]: I1008 15:01:01.831661 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332261-6nq76" event={"ID":"caa89bd3-cdc8-412a-82f0-1289cf0028c1","Type":"ContainerStarted","Data":"e9afb4b4412e71700d63f5f4c9bf54fef961aad1181262f87532ff0cf327247c"} Oct 08 15:01:01 crc kubenswrapper[4789]: I1008 15:01:01.857470 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h47hs" podStartSLOduration=3.178689114 podStartE2EDuration="7.857447768s" podCreationTimestamp="2025-10-08 15:00:54 +0000 UTC" firstStartedPulling="2025-10-08 15:00:56.764514294 +0000 UTC m=+3596.671261786" lastFinishedPulling="2025-10-08 15:01:01.443272948 +0000 UTC m=+3601.350020440" observedRunningTime="2025-10-08 15:01:01.849880921 +0000 UTC m=+3601.756628413" watchObservedRunningTime="2025-10-08 15:01:01.857447768 +0000 UTC m=+3601.764195260" Oct 08 15:01:01 crc kubenswrapper[4789]: I1008 15:01:01.876511 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29332261-6nq76" podStartSLOduration=1.876489207 podStartE2EDuration="1.876489207s" podCreationTimestamp="2025-10-08 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:01:01.864754707 +0000 UTC m=+3601.771502199" watchObservedRunningTime="2025-10-08 15:01:01.876489207 +0000 UTC m=+3601.783236699" Oct 08 15:01:05 crc kubenswrapper[4789]: I1008 15:01:05.123883 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:01:05 crc kubenswrapper[4789]: I1008 15:01:05.124456 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:01:05 crc kubenswrapper[4789]: I1008 15:01:05.191387 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:01:05 crc kubenswrapper[4789]: I1008 15:01:05.876214 4789 generic.go:334] "Generic (PLEG): container finished" podID="caa89bd3-cdc8-412a-82f0-1289cf0028c1" containerID="5be03d1c82ee66b2fdad68abf2d4e2a2369013da360b5c7cb40b0330b039185e" exitCode=0 Oct 08 15:01:05 crc kubenswrapper[4789]: I1008 15:01:05.876447 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332261-6nq76" event={"ID":"caa89bd3-cdc8-412a-82f0-1289cf0028c1","Type":"ContainerDied","Data":"5be03d1c82ee66b2fdad68abf2d4e2a2369013da360b5c7cb40b0330b039185e"} Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.286111 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.452120 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-combined-ca-bundle\") pod \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.452187 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-fernet-keys\") pod \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.452465 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-config-data\") pod \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.452575 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rw6v\" (UniqueName: \"kubernetes.io/projected/caa89bd3-cdc8-412a-82f0-1289cf0028c1-kube-api-access-6rw6v\") pod \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\" (UID: \"caa89bd3-cdc8-412a-82f0-1289cf0028c1\") " Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.458153 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa89bd3-cdc8-412a-82f0-1289cf0028c1-kube-api-access-6rw6v" (OuterVolumeSpecName: "kube-api-access-6rw6v") pod "caa89bd3-cdc8-412a-82f0-1289cf0028c1" (UID: "caa89bd3-cdc8-412a-82f0-1289cf0028c1"). InnerVolumeSpecName "kube-api-access-6rw6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.472427 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "caa89bd3-cdc8-412a-82f0-1289cf0028c1" (UID: "caa89bd3-cdc8-412a-82f0-1289cf0028c1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.482785 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "caa89bd3-cdc8-412a-82f0-1289cf0028c1" (UID: "caa89bd3-cdc8-412a-82f0-1289cf0028c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.538178 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-config-data" (OuterVolumeSpecName: "config-data") pod "caa89bd3-cdc8-412a-82f0-1289cf0028c1" (UID: "caa89bd3-cdc8-412a-82f0-1289cf0028c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.556004 4789 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.556086 4789 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.556100 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa89bd3-cdc8-412a-82f0-1289cf0028c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.556113 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rw6v\" (UniqueName: \"kubernetes.io/projected/caa89bd3-cdc8-412a-82f0-1289cf0028c1-kube-api-access-6rw6v\") on node \"crc\" DevicePath \"\"" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.902851 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332261-6nq76" event={"ID":"caa89bd3-cdc8-412a-82f0-1289cf0028c1","Type":"ContainerDied","Data":"e9afb4b4412e71700d63f5f4c9bf54fef961aad1181262f87532ff0cf327247c"} Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.903151 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9afb4b4412e71700d63f5f4c9bf54fef961aad1181262f87532ff0cf327247c" Oct 08 15:01:07 crc kubenswrapper[4789]: I1008 15:01:07.902912 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332261-6nq76" Oct 08 15:01:15 crc kubenswrapper[4789]: I1008 15:01:15.197831 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:01:15 crc kubenswrapper[4789]: I1008 15:01:15.307864 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h47hs"] Oct 08 15:01:15 crc kubenswrapper[4789]: I1008 15:01:15.978797 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h47hs" podUID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerName="registry-server" containerID="cri-o://270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f" gracePeriod=2 Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.470325 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.660173 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6m46\" (UniqueName: \"kubernetes.io/projected/4b5bc008-5174-4e13-a71f-be5f9bd756cf-kube-api-access-x6m46\") pod \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.660531 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-catalog-content\") pod \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.660604 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-utilities\") pod \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\" (UID: \"4b5bc008-5174-4e13-a71f-be5f9bd756cf\") " Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.661366 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-utilities" (OuterVolumeSpecName: "utilities") pod "4b5bc008-5174-4e13-a71f-be5f9bd756cf" (UID: "4b5bc008-5174-4e13-a71f-be5f9bd756cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.680774 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b5bc008-5174-4e13-a71f-be5f9bd756cf-kube-api-access-x6m46" (OuterVolumeSpecName: "kube-api-access-x6m46") pod "4b5bc008-5174-4e13-a71f-be5f9bd756cf" (UID: "4b5bc008-5174-4e13-a71f-be5f9bd756cf"). InnerVolumeSpecName "kube-api-access-x6m46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.705509 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b5bc008-5174-4e13-a71f-be5f9bd756cf" (UID: "4b5bc008-5174-4e13-a71f-be5f9bd756cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.763051 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6m46\" (UniqueName: \"kubernetes.io/projected/4b5bc008-5174-4e13-a71f-be5f9bd756cf-kube-api-access-x6m46\") on node \"crc\" DevicePath \"\"" Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.763086 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.763095 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b5bc008-5174-4e13-a71f-be5f9bd756cf-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.990004 4789 generic.go:334] "Generic (PLEG): container finished" podID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerID="270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f" exitCode=0 Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.990052 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47hs" event={"ID":"4b5bc008-5174-4e13-a71f-be5f9bd756cf","Type":"ContainerDied","Data":"270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f"} Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.990088 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h47hs" event={"ID":"4b5bc008-5174-4e13-a71f-be5f9bd756cf","Type":"ContainerDied","Data":"760f9fdf9b79fc6976ffa56d5bf266c4cea2aacfd573b1bdbe7713de3ddf7271"} Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.990113 4789 scope.go:117] "RemoveContainer" containerID="270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f" Oct 08 15:01:16 crc kubenswrapper[4789]: I1008 15:01:16.990111 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h47hs" Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.020178 4789 scope.go:117] "RemoveContainer" containerID="6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7" Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.022618 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h47hs"] Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.034046 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h47hs"] Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.041852 4789 scope.go:117] "RemoveContainer" containerID="968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde" Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.087504 4789 scope.go:117] "RemoveContainer" containerID="270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f" Oct 08 15:01:17 crc kubenswrapper[4789]: E1008 15:01:17.087914 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f\": container with ID starting with 270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f not found: ID does not exist" containerID="270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f" Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.087973 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f"} err="failed to get container status \"270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f\": rpc error: code = NotFound desc = could not find container \"270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f\": container with ID starting with 270e372a44003af46ab4721482e05d9e202a64f4b0916618f414912a31a3113f not found: ID does not exist" Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.088025 4789 scope.go:117] "RemoveContainer" containerID="6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7" Oct 08 15:01:17 crc kubenswrapper[4789]: E1008 15:01:17.088396 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7\": container with ID starting with 6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7 not found: ID does not exist" containerID="6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7" Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.088429 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7"} err="failed to get container status \"6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7\": rpc error: code = NotFound desc = could not find container \"6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7\": container with ID starting with 6d8ec001949c8432ff87197ad6670325bcdf2184001480f19d8e06820d18b6c7 not found: ID does not exist" Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.088453 4789 scope.go:117] "RemoveContainer" containerID="968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde" Oct 08 15:01:17 crc kubenswrapper[4789]: E1008 15:01:17.088769 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde\": container with ID starting with 968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde not found: ID does not exist" containerID="968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde" Oct 08 15:01:17 crc kubenswrapper[4789]: I1008 15:01:17.088788 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde"} err="failed to get container status \"968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde\": rpc error: code = NotFound desc = could not find container \"968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde\": container with ID starting with 968a1ebdcf53b88e8b99266252c2572a2d716e71fc30214143ca1933908bbcde not found: ID does not exist" Oct 08 15:01:18 crc kubenswrapper[4789]: I1008 15:01:18.745406 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" path="/var/lib/kubelet/pods/4b5bc008-5174-4e13-a71f-be5f9bd756cf/volumes" Oct 08 15:01:26 crc kubenswrapper[4789]: I1008 15:01:26.433039 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:01:26 crc kubenswrapper[4789]: I1008 15:01:26.433623 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:01:56 crc kubenswrapper[4789]: I1008 15:01:56.432932 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:01:56 crc kubenswrapper[4789]: I1008 15:01:56.433684 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:01:56 crc kubenswrapper[4789]: I1008 15:01:56.433743 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 15:01:56 crc kubenswrapper[4789]: I1008 15:01:56.434688 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:01:56 crc kubenswrapper[4789]: I1008 15:01:56.434747 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" gracePeriod=600 Oct 08 15:01:56 crc kubenswrapper[4789]: E1008 15:01:56.562841 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:01:57 crc kubenswrapper[4789]: I1008 15:01:57.412137 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" exitCode=0 Oct 08 15:01:57 crc kubenswrapper[4789]: I1008 15:01:57.412231 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea"} Oct 08 15:01:57 crc kubenswrapper[4789]: I1008 15:01:57.412548 4789 scope.go:117] "RemoveContainer" containerID="03559d7c0bfab1b8eaf3db8cb07705d99d1ef3a70471e25a70558c926b95793c" Oct 08 15:01:57 crc kubenswrapper[4789]: I1008 15:01:57.414280 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:01:57 crc kubenswrapper[4789]: E1008 15:01:57.415216 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:02:08 crc kubenswrapper[4789]: I1008 15:02:08.730547 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:02:08 crc kubenswrapper[4789]: E1008 15:02:08.731264 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:02:21 crc kubenswrapper[4789]: I1008 15:02:21.730314 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:02:21 crc kubenswrapper[4789]: E1008 15:02:21.731085 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:02:35 crc kubenswrapper[4789]: I1008 15:02:35.731633 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:02:35 crc kubenswrapper[4789]: E1008 15:02:35.732474 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:02:36 crc kubenswrapper[4789]: E1008 15:02:36.890768 4789 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 08 15:02:48 crc kubenswrapper[4789]: I1008 15:02:48.730946 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:02:48 crc kubenswrapper[4789]: E1008 15:02:48.733393 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:03:00 crc kubenswrapper[4789]: I1008 15:03:00.746491 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:03:00 crc kubenswrapper[4789]: E1008 15:03:00.747767 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:03:15 crc kubenswrapper[4789]: I1008 15:03:15.730361 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:03:15 crc kubenswrapper[4789]: E1008 15:03:15.731228 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:03:24 crc kubenswrapper[4789]: I1008 15:03:24.271062 4789 scope.go:117] "RemoveContainer" containerID="a57e8a500c230d6969337f6f8a9ea518499c5f1bb0f4238009dd18a41b64e144" Oct 08 15:03:24 crc kubenswrapper[4789]: I1008 15:03:24.301922 4789 scope.go:117] "RemoveContainer" containerID="dec10a416c9e1898d2aa7a401ccff0d83174008ee344a43e2c08c06d014dc46c" Oct 08 15:03:24 crc kubenswrapper[4789]: I1008 15:03:24.344937 4789 scope.go:117] "RemoveContainer" containerID="a929595bebde030e07ee01ffaafc1d0f14a6ca599c11425f696f7618ffef57bc" Oct 08 15:03:27 crc kubenswrapper[4789]: I1008 15:03:27.730266 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:03:27 crc kubenswrapper[4789]: E1008 15:03:27.730567 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:03:41 crc kubenswrapper[4789]: I1008 15:03:41.730634 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:03:41 crc kubenswrapper[4789]: E1008 15:03:41.731623 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:03:53 crc kubenswrapper[4789]: I1008 15:03:53.738582 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:03:53 crc kubenswrapper[4789]: E1008 15:03:53.739674 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:04:06 crc kubenswrapper[4789]: I1008 15:04:06.731193 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:04:06 crc kubenswrapper[4789]: E1008 15:04:06.732033 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:04:20 crc kubenswrapper[4789]: I1008 15:04:20.739870 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:04:20 crc kubenswrapper[4789]: E1008 15:04:20.740579 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:04:33 crc kubenswrapper[4789]: I1008 15:04:33.730354 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:04:33 crc kubenswrapper[4789]: E1008 15:04:33.732388 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:04:44 crc kubenswrapper[4789]: I1008 15:04:44.730485 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:04:44 crc kubenswrapper[4789]: E1008 15:04:44.731419 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:04:58 crc kubenswrapper[4789]: I1008 15:04:58.731829 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:04:58 crc kubenswrapper[4789]: E1008 15:04:58.733097 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:05:12 crc kubenswrapper[4789]: I1008 15:05:12.733336 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:05:12 crc kubenswrapper[4789]: E1008 15:05:12.734247 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:05:25 crc kubenswrapper[4789]: I1008 15:05:25.752319 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:05:25 crc kubenswrapper[4789]: E1008 15:05:25.753110 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.453168 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-585ll"] Oct 08 15:05:27 crc kubenswrapper[4789]: E1008 15:05:27.454804 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerName="registry-server" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.454842 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerName="registry-server" Oct 08 15:05:27 crc kubenswrapper[4789]: E1008 15:05:27.454910 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerName="extract-content" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.454935 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerName="extract-content" Oct 08 15:05:27 crc kubenswrapper[4789]: E1008 15:05:27.454956 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa89bd3-cdc8-412a-82f0-1289cf0028c1" containerName="keystone-cron" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.455028 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa89bd3-cdc8-412a-82f0-1289cf0028c1" containerName="keystone-cron" Oct 08 15:05:27 crc kubenswrapper[4789]: E1008 15:05:27.455075 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerName="extract-utilities" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.455083 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerName="extract-utilities" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.455423 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa89bd3-cdc8-412a-82f0-1289cf0028c1" containerName="keystone-cron" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.455477 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b5bc008-5174-4e13-a71f-be5f9bd756cf" containerName="registry-server" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.458760 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.462459 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-585ll"] Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.546418 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-catalog-content\") pod \"redhat-operators-585ll\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.546548 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-utilities\") pod \"redhat-operators-585ll\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.546601 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4wss\" (UniqueName: \"kubernetes.io/projected/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-kube-api-access-p4wss\") pod \"redhat-operators-585ll\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.648308 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-utilities\") pod \"redhat-operators-585ll\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.648399 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4wss\" (UniqueName: \"kubernetes.io/projected/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-kube-api-access-p4wss\") pod \"redhat-operators-585ll\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.648500 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-catalog-content\") pod \"redhat-operators-585ll\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.648884 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-utilities\") pod \"redhat-operators-585ll\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.648949 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-catalog-content\") pod \"redhat-operators-585ll\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.679912 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4wss\" (UniqueName: \"kubernetes.io/projected/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-kube-api-access-p4wss\") pod \"redhat-operators-585ll\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:27 crc kubenswrapper[4789]: I1008 15:05:27.779032 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:28 crc kubenswrapper[4789]: I1008 15:05:28.312918 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-585ll"] Oct 08 15:05:28 crc kubenswrapper[4789]: I1008 15:05:28.499857 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-585ll" event={"ID":"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272","Type":"ContainerStarted","Data":"ba35bc2901a6b705bb21c744253fdfe69daf46e645f0d2fdeccaf5dfbf5940d7"} Oct 08 15:05:29 crc kubenswrapper[4789]: I1008 15:05:29.513391 4789 generic.go:334] "Generic (PLEG): container finished" podID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerID="c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837" exitCode=0 Oct 08 15:05:29 crc kubenswrapper[4789]: I1008 15:05:29.513499 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-585ll" event={"ID":"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272","Type":"ContainerDied","Data":"c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837"} Oct 08 15:05:31 crc kubenswrapper[4789]: I1008 15:05:31.533144 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-585ll" event={"ID":"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272","Type":"ContainerStarted","Data":"2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59"} Oct 08 15:05:37 crc kubenswrapper[4789]: I1008 15:05:37.588864 4789 generic.go:334] "Generic (PLEG): container finished" podID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerID="2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59" exitCode=0 Oct 08 15:05:37 crc kubenswrapper[4789]: I1008 15:05:37.588929 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-585ll" event={"ID":"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272","Type":"ContainerDied","Data":"2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59"} Oct 08 15:05:37 crc kubenswrapper[4789]: I1008 15:05:37.730020 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:05:37 crc kubenswrapper[4789]: E1008 15:05:37.730271 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:05:38 crc kubenswrapper[4789]: I1008 15:05:38.599606 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-585ll" event={"ID":"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272","Type":"ContainerStarted","Data":"9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4"} Oct 08 15:05:38 crc kubenswrapper[4789]: I1008 15:05:38.615221 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-585ll" podStartSLOduration=2.928338901 podStartE2EDuration="11.615206057s" podCreationTimestamp="2025-10-08 15:05:27 +0000 UTC" firstStartedPulling="2025-10-08 15:05:29.515298604 +0000 UTC m=+3869.422046096" lastFinishedPulling="2025-10-08 15:05:38.20216576 +0000 UTC m=+3878.108913252" observedRunningTime="2025-10-08 15:05:38.613709086 +0000 UTC m=+3878.520456578" watchObservedRunningTime="2025-10-08 15:05:38.615206057 +0000 UTC m=+3878.521953549" Oct 08 15:05:47 crc kubenswrapper[4789]: I1008 15:05:47.779855 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:47 crc kubenswrapper[4789]: I1008 15:05:47.780408 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:05:48 crc kubenswrapper[4789]: I1008 15:05:48.826337 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-585ll" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="registry-server" probeResult="failure" output=< Oct 08 15:05:48 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 15:05:48 crc kubenswrapper[4789]: > Oct 08 15:05:51 crc kubenswrapper[4789]: I1008 15:05:51.730410 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:05:51 crc kubenswrapper[4789]: E1008 15:05:51.731455 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:05:58 crc kubenswrapper[4789]: I1008 15:05:58.823084 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-585ll" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="registry-server" probeResult="failure" output=< Oct 08 15:05:58 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 15:05:58 crc kubenswrapper[4789]: > Oct 08 15:06:05 crc kubenswrapper[4789]: I1008 15:06:05.730093 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:06:05 crc kubenswrapper[4789]: E1008 15:06:05.730775 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:06:07 crc kubenswrapper[4789]: I1008 15:06:07.832537 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:06:07 crc kubenswrapper[4789]: I1008 15:06:07.885521 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:06:08 crc kubenswrapper[4789]: I1008 15:06:08.064785 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-585ll"] Oct 08 15:06:08 crc kubenswrapper[4789]: I1008 15:06:08.876704 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-585ll" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="registry-server" containerID="cri-o://9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4" gracePeriod=2 Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.411367 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.520621 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-utilities\") pod \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.520704 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-catalog-content\") pod \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.520762 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4wss\" (UniqueName: \"kubernetes.io/projected/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-kube-api-access-p4wss\") pod \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\" (UID: \"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272\") " Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.521555 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-utilities" (OuterVolumeSpecName: "utilities") pod "8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" (UID: "8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.532221 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-kube-api-access-p4wss" (OuterVolumeSpecName: "kube-api-access-p4wss") pod "8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" (UID: "8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272"). InnerVolumeSpecName "kube-api-access-p4wss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.609705 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" (UID: "8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.623290 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.623327 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.623339 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4wss\" (UniqueName: \"kubernetes.io/projected/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272-kube-api-access-p4wss\") on node \"crc\" DevicePath \"\"" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.895917 4789 generic.go:334] "Generic (PLEG): container finished" podID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerID="9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4" exitCode=0 Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.895961 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-585ll" event={"ID":"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272","Type":"ContainerDied","Data":"9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4"} Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.896009 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-585ll" event={"ID":"8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272","Type":"ContainerDied","Data":"ba35bc2901a6b705bb21c744253fdfe69daf46e645f0d2fdeccaf5dfbf5940d7"} Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.896028 4789 scope.go:117] "RemoveContainer" containerID="9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.896077 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-585ll" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.918127 4789 scope.go:117] "RemoveContainer" containerID="2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.960707 4789 scope.go:117] "RemoveContainer" containerID="c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837" Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.966053 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-585ll"] Oct 08 15:06:09 crc kubenswrapper[4789]: I1008 15:06:09.979895 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-585ll"] Oct 08 15:06:10 crc kubenswrapper[4789]: I1008 15:06:10.004815 4789 scope.go:117] "RemoveContainer" containerID="9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4" Oct 08 15:06:10 crc kubenswrapper[4789]: E1008 15:06:10.005293 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4\": container with ID starting with 9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4 not found: ID does not exist" containerID="9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4" Oct 08 15:06:10 crc kubenswrapper[4789]: I1008 15:06:10.005349 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4"} err="failed to get container status \"9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4\": rpc error: code = NotFound desc = could not find container \"9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4\": container with ID starting with 9690f87f4f02079bc657040278cd6e9a7b0d12fc8e6dd7ba276ffae1c8ea02e4 not found: ID does not exist" Oct 08 15:06:10 crc kubenswrapper[4789]: I1008 15:06:10.005416 4789 scope.go:117] "RemoveContainer" containerID="2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59" Oct 08 15:06:10 crc kubenswrapper[4789]: E1008 15:06:10.005722 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59\": container with ID starting with 2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59 not found: ID does not exist" containerID="2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59" Oct 08 15:06:10 crc kubenswrapper[4789]: I1008 15:06:10.005760 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59"} err="failed to get container status \"2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59\": rpc error: code = NotFound desc = could not find container \"2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59\": container with ID starting with 2f679175fe2a47065909ec483ccee73f9bb5f92f08f65ae3abff75c535a69a59 not found: ID does not exist" Oct 08 15:06:10 crc kubenswrapper[4789]: I1008 15:06:10.005785 4789 scope.go:117] "RemoveContainer" containerID="c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837" Oct 08 15:06:10 crc kubenswrapper[4789]: E1008 15:06:10.006020 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837\": container with ID starting with c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837 not found: ID does not exist" containerID="c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837" Oct 08 15:06:10 crc kubenswrapper[4789]: I1008 15:06:10.006045 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837"} err="failed to get container status \"c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837\": rpc error: code = NotFound desc = could not find container \"c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837\": container with ID starting with c4d09b1840a466a6d3fec11f5a60302683bd8f20c66d443d5a4f5a913fad2837 not found: ID does not exist" Oct 08 15:06:10 crc kubenswrapper[4789]: I1008 15:06:10.742410 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" path="/var/lib/kubelet/pods/8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272/volumes" Oct 08 15:06:18 crc kubenswrapper[4789]: I1008 15:06:18.731227 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:06:18 crc kubenswrapper[4789]: E1008 15:06:18.732144 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:06:33 crc kubenswrapper[4789]: I1008 15:06:33.729744 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:06:33 crc kubenswrapper[4789]: E1008 15:06:33.730522 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:06:46 crc kubenswrapper[4789]: I1008 15:06:46.731393 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:06:46 crc kubenswrapper[4789]: E1008 15:06:46.732262 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:07:00 crc kubenswrapper[4789]: I1008 15:07:00.740262 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:07:01 crc kubenswrapper[4789]: I1008 15:07:01.438978 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"73a99c9d7a580b4700b76660ad13798cfcc0758d3e56372942489377a92f0c2c"} Oct 08 15:09:26 crc kubenswrapper[4789]: I1008 15:09:26.433409 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:09:26 crc kubenswrapper[4789]: I1008 15:09:26.434529 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.791111 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p54ww"] Oct 08 15:09:39 crc kubenswrapper[4789]: E1008 15:09:39.792055 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="extract-utilities" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.792070 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="extract-utilities" Oct 08 15:09:39 crc kubenswrapper[4789]: E1008 15:09:39.792087 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="registry-server" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.792093 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="registry-server" Oct 08 15:09:39 crc kubenswrapper[4789]: E1008 15:09:39.792107 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="extract-content" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.792115 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="extract-content" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.792446 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9f7624-bfe4-4aaa-9e0d-d8bf0ed28272" containerName="registry-server" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.794078 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.819441 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p54ww"] Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.825061 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9z5w\" (UniqueName: \"kubernetes.io/projected/91f9b642-b86f-4815-b903-e74f1e17dc4d-kube-api-access-j9z5w\") pod \"community-operators-p54ww\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.825110 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-catalog-content\") pod \"community-operators-p54ww\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.825514 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-utilities\") pod \"community-operators-p54ww\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.927192 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9z5w\" (UniqueName: \"kubernetes.io/projected/91f9b642-b86f-4815-b903-e74f1e17dc4d-kube-api-access-j9z5w\") pod \"community-operators-p54ww\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.927246 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-catalog-content\") pod \"community-operators-p54ww\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.927277 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-utilities\") pod \"community-operators-p54ww\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.927711 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-utilities\") pod \"community-operators-p54ww\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.927819 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-catalog-content\") pod \"community-operators-p54ww\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:39 crc kubenswrapper[4789]: I1008 15:09:39.970041 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9z5w\" (UniqueName: \"kubernetes.io/projected/91f9b642-b86f-4815-b903-e74f1e17dc4d-kube-api-access-j9z5w\") pod \"community-operators-p54ww\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:40 crc kubenswrapper[4789]: I1008 15:09:40.119155 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:40 crc kubenswrapper[4789]: I1008 15:09:40.605461 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p54ww"] Oct 08 15:09:40 crc kubenswrapper[4789]: W1008 15:09:40.613456 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91f9b642_b86f_4815_b903_e74f1e17dc4d.slice/crio-b6c7d1b4c21a1779f48d66bc9e2e48809da4e09d61fc3db08b5aca861178e2cb WatchSource:0}: Error finding container b6c7d1b4c21a1779f48d66bc9e2e48809da4e09d61fc3db08b5aca861178e2cb: Status 404 returned error can't find the container with id b6c7d1b4c21a1779f48d66bc9e2e48809da4e09d61fc3db08b5aca861178e2cb Oct 08 15:09:41 crc kubenswrapper[4789]: I1008 15:09:41.015634 4789 generic.go:334] "Generic (PLEG): container finished" podID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerID="bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b" exitCode=0 Oct 08 15:09:41 crc kubenswrapper[4789]: I1008 15:09:41.015694 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p54ww" event={"ID":"91f9b642-b86f-4815-b903-e74f1e17dc4d","Type":"ContainerDied","Data":"bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b"} Oct 08 15:09:41 crc kubenswrapper[4789]: I1008 15:09:41.015757 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p54ww" event={"ID":"91f9b642-b86f-4815-b903-e74f1e17dc4d","Type":"ContainerStarted","Data":"b6c7d1b4c21a1779f48d66bc9e2e48809da4e09d61fc3db08b5aca861178e2cb"} Oct 08 15:09:41 crc kubenswrapper[4789]: I1008 15:09:41.019905 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:09:43 crc kubenswrapper[4789]: I1008 15:09:43.036306 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p54ww" event={"ID":"91f9b642-b86f-4815-b903-e74f1e17dc4d","Type":"ContainerStarted","Data":"3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f"} Oct 08 15:09:43 crc kubenswrapper[4789]: I1008 15:09:43.787238 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8x8wr"] Oct 08 15:09:43 crc kubenswrapper[4789]: I1008 15:09:43.789893 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:43 crc kubenswrapper[4789]: I1008 15:09:43.799481 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8x8wr"] Oct 08 15:09:43 crc kubenswrapper[4789]: I1008 15:09:43.918540 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-utilities\") pod \"redhat-marketplace-8x8wr\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:43 crc kubenswrapper[4789]: I1008 15:09:43.918946 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-catalog-content\") pod \"redhat-marketplace-8x8wr\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:43 crc kubenswrapper[4789]: I1008 15:09:43.919128 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngcfz\" (UniqueName: \"kubernetes.io/projected/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-kube-api-access-ngcfz\") pod \"redhat-marketplace-8x8wr\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.020946 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngcfz\" (UniqueName: \"kubernetes.io/projected/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-kube-api-access-ngcfz\") pod \"redhat-marketplace-8x8wr\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.021102 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-utilities\") pod \"redhat-marketplace-8x8wr\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.021203 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-catalog-content\") pod \"redhat-marketplace-8x8wr\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.021640 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-catalog-content\") pod \"redhat-marketplace-8x8wr\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.022249 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-utilities\") pod \"redhat-marketplace-8x8wr\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.045396 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngcfz\" (UniqueName: \"kubernetes.io/projected/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-kube-api-access-ngcfz\") pod \"redhat-marketplace-8x8wr\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.048360 4789 generic.go:334] "Generic (PLEG): container finished" podID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerID="3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f" exitCode=0 Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.048395 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p54ww" event={"ID":"91f9b642-b86f-4815-b903-e74f1e17dc4d","Type":"ContainerDied","Data":"3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f"} Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.113545 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:09:44 crc kubenswrapper[4789]: I1008 15:09:44.587953 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8x8wr"] Oct 08 15:09:44 crc kubenswrapper[4789]: W1008 15:09:44.588027 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30fcfaf9_6a90_4e41_9d64_6fb9b1a310a0.slice/crio-2758136f538c0e282f3f5f4e1b2e49b46d65b937d1ae1a89583ea7509019ef37 WatchSource:0}: Error finding container 2758136f538c0e282f3f5f4e1b2e49b46d65b937d1ae1a89583ea7509019ef37: Status 404 returned error can't find the container with id 2758136f538c0e282f3f5f4e1b2e49b46d65b937d1ae1a89583ea7509019ef37 Oct 08 15:09:45 crc kubenswrapper[4789]: I1008 15:09:45.057832 4789 generic.go:334] "Generic (PLEG): container finished" podID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerID="6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7" exitCode=0 Oct 08 15:09:45 crc kubenswrapper[4789]: I1008 15:09:45.058268 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8x8wr" event={"ID":"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0","Type":"ContainerDied","Data":"6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7"} Oct 08 15:09:45 crc kubenswrapper[4789]: I1008 15:09:45.058293 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8x8wr" event={"ID":"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0","Type":"ContainerStarted","Data":"2758136f538c0e282f3f5f4e1b2e49b46d65b937d1ae1a89583ea7509019ef37"} Oct 08 15:09:45 crc kubenswrapper[4789]: I1008 15:09:45.062724 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p54ww" event={"ID":"91f9b642-b86f-4815-b903-e74f1e17dc4d","Type":"ContainerStarted","Data":"60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022"} Oct 08 15:09:45 crc kubenswrapper[4789]: I1008 15:09:45.108084 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p54ww" podStartSLOduration=2.627550245 podStartE2EDuration="6.108062981s" podCreationTimestamp="2025-10-08 15:09:39 +0000 UTC" firstStartedPulling="2025-10-08 15:09:41.019668587 +0000 UTC m=+4120.926416079" lastFinishedPulling="2025-10-08 15:09:44.500181323 +0000 UTC m=+4124.406928815" observedRunningTime="2025-10-08 15:09:45.098094789 +0000 UTC m=+4125.004842281" watchObservedRunningTime="2025-10-08 15:09:45.108062981 +0000 UTC m=+4125.014810483" Oct 08 15:09:49 crc kubenswrapper[4789]: I1008 15:09:49.103227 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8x8wr" event={"ID":"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0","Type":"ContainerStarted","Data":"5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc"} Oct 08 15:09:50 crc kubenswrapper[4789]: I1008 15:09:50.116036 4789 generic.go:334] "Generic (PLEG): container finished" podID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerID="5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc" exitCode=0 Oct 08 15:09:50 crc kubenswrapper[4789]: I1008 15:09:50.116110 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8x8wr" event={"ID":"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0","Type":"ContainerDied","Data":"5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc"} Oct 08 15:09:50 crc kubenswrapper[4789]: I1008 15:09:50.120276 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:50 crc kubenswrapper[4789]: I1008 15:09:50.120339 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:50 crc kubenswrapper[4789]: I1008 15:09:50.192499 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:51 crc kubenswrapper[4789]: I1008 15:09:51.191716 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:52 crc kubenswrapper[4789]: I1008 15:09:52.777647 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p54ww"] Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.170146 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8x8wr" event={"ID":"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0","Type":"ContainerStarted","Data":"3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6"} Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.170659 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p54ww" podUID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerName="registry-server" containerID="cri-o://60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022" gracePeriod=2 Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.201496 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8x8wr" podStartSLOduration=2.829643447 podStartE2EDuration="11.201471768s" podCreationTimestamp="2025-10-08 15:09:43 +0000 UTC" firstStartedPulling="2025-10-08 15:09:45.059718043 +0000 UTC m=+4124.966465535" lastFinishedPulling="2025-10-08 15:09:53.431546354 +0000 UTC m=+4133.338293856" observedRunningTime="2025-10-08 15:09:54.191226269 +0000 UTC m=+4134.097973771" watchObservedRunningTime="2025-10-08 15:09:54.201471768 +0000 UTC m=+4134.108219260" Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.766029 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.846933 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-catalog-content\") pod \"91f9b642-b86f-4815-b903-e74f1e17dc4d\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.847022 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9z5w\" (UniqueName: \"kubernetes.io/projected/91f9b642-b86f-4815-b903-e74f1e17dc4d-kube-api-access-j9z5w\") pod \"91f9b642-b86f-4815-b903-e74f1e17dc4d\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.847159 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-utilities\") pod \"91f9b642-b86f-4815-b903-e74f1e17dc4d\" (UID: \"91f9b642-b86f-4815-b903-e74f1e17dc4d\") " Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.848074 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-utilities" (OuterVolumeSpecName: "utilities") pod "91f9b642-b86f-4815-b903-e74f1e17dc4d" (UID: "91f9b642-b86f-4815-b903-e74f1e17dc4d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.849486 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.860265 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f9b642-b86f-4815-b903-e74f1e17dc4d-kube-api-access-j9z5w" (OuterVolumeSpecName: "kube-api-access-j9z5w") pod "91f9b642-b86f-4815-b903-e74f1e17dc4d" (UID: "91f9b642-b86f-4815-b903-e74f1e17dc4d"). InnerVolumeSpecName "kube-api-access-j9z5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.950871 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9z5w\" (UniqueName: \"kubernetes.io/projected/91f9b642-b86f-4815-b903-e74f1e17dc4d-kube-api-access-j9z5w\") on node \"crc\" DevicePath \"\"" Oct 08 15:09:54 crc kubenswrapper[4789]: I1008 15:09:54.974930 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91f9b642-b86f-4815-b903-e74f1e17dc4d" (UID: "91f9b642-b86f-4815-b903-e74f1e17dc4d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:09:55 crc kubenswrapper[4789]: I1008 15:09:55.052759 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f9b642-b86f-4815-b903-e74f1e17dc4d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:09:55 crc kubenswrapper[4789]: I1008 15:09:55.185772 4789 generic.go:334] "Generic (PLEG): container finished" podID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerID="60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022" exitCode=0 Oct 08 15:09:55 crc kubenswrapper[4789]: I1008 15:09:55.185851 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p54ww" Oct 08 15:09:55 crc kubenswrapper[4789]: I1008 15:09:55.185891 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p54ww" event={"ID":"91f9b642-b86f-4815-b903-e74f1e17dc4d","Type":"ContainerDied","Data":"60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022"} Oct 08 15:09:55 crc kubenswrapper[4789]: I1008 15:09:55.185972 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p54ww" event={"ID":"91f9b642-b86f-4815-b903-e74f1e17dc4d","Type":"ContainerDied","Data":"b6c7d1b4c21a1779f48d66bc9e2e48809da4e09d61fc3db08b5aca861178e2cb"} Oct 08 15:09:55 crc kubenswrapper[4789]: I1008 15:09:55.186047 4789 scope.go:117] "RemoveContainer" containerID="60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022" Oct 08 15:09:55 crc kubenswrapper[4789]: I1008 15:09:55.220952 4789 scope.go:117] "RemoveContainer" containerID="3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f" Oct 08 15:09:55 crc kubenswrapper[4789]: I1008 15:09:55.250300 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p54ww"] Oct 08 15:09:55 crc kubenswrapper[4789]: I1008 15:09:55.285099 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p54ww"] Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.096577 4789 scope.go:117] "RemoveContainer" containerID="bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b" Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.260410 4789 scope.go:117] "RemoveContainer" containerID="60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022" Oct 08 15:09:56 crc kubenswrapper[4789]: E1008 15:09:56.261178 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022\": container with ID starting with 60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022 not found: ID does not exist" containerID="60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022" Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.261221 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022"} err="failed to get container status \"60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022\": rpc error: code = NotFound desc = could not find container \"60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022\": container with ID starting with 60931798fe15e87433cd514ce06d2f03f62294b5af6883352477ed2081922022 not found: ID does not exist" Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.261250 4789 scope.go:117] "RemoveContainer" containerID="3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f" Oct 08 15:09:56 crc kubenswrapper[4789]: E1008 15:09:56.261608 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f\": container with ID starting with 3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f not found: ID does not exist" containerID="3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f" Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.261685 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f"} err="failed to get container status \"3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f\": rpc error: code = NotFound desc = could not find container \"3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f\": container with ID starting with 3362e7905635c216d299790679ae1b90912b3b4eafd85d75a0377bc2c590d94f not found: ID does not exist" Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.261734 4789 scope.go:117] "RemoveContainer" containerID="bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b" Oct 08 15:09:56 crc kubenswrapper[4789]: E1008 15:09:56.262439 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b\": container with ID starting with bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b not found: ID does not exist" containerID="bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b" Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.262469 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b"} err="failed to get container status \"bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b\": rpc error: code = NotFound desc = could not find container \"bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b\": container with ID starting with bc7ae13261ee36ffce77891b0df56e055b37afd6586ac81adf5e58039862ae3b not found: ID does not exist" Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.433724 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.433796 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:09:56 crc kubenswrapper[4789]: I1008 15:09:56.746659 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f9b642-b86f-4815-b903-e74f1e17dc4d" path="/var/lib/kubelet/pods/91f9b642-b86f-4815-b903-e74f1e17dc4d/volumes" Oct 08 15:10:04 crc kubenswrapper[4789]: I1008 15:10:04.114931 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:10:04 crc kubenswrapper[4789]: I1008 15:10:04.116618 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:10:04 crc kubenswrapper[4789]: I1008 15:10:04.168721 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:10:04 crc kubenswrapper[4789]: I1008 15:10:04.328501 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:10:04 crc kubenswrapper[4789]: I1008 15:10:04.401903 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8x8wr"] Oct 08 15:10:06 crc kubenswrapper[4789]: I1008 15:10:06.303460 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8x8wr" podUID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerName="registry-server" containerID="cri-o://3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6" gracePeriod=2 Oct 08 15:10:06 crc kubenswrapper[4789]: I1008 15:10:06.783892 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:10:06 crc kubenswrapper[4789]: I1008 15:10:06.945253 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-catalog-content\") pod \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " Oct 08 15:10:06 crc kubenswrapper[4789]: I1008 15:10:06.945714 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngcfz\" (UniqueName: \"kubernetes.io/projected/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-kube-api-access-ngcfz\") pod \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " Oct 08 15:10:06 crc kubenswrapper[4789]: I1008 15:10:06.945793 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-utilities\") pod \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\" (UID: \"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0\") " Oct 08 15:10:06 crc kubenswrapper[4789]: I1008 15:10:06.946799 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-utilities" (OuterVolumeSpecName: "utilities") pod "30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" (UID: "30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:10:06 crc kubenswrapper[4789]: I1008 15:10:06.953361 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-kube-api-access-ngcfz" (OuterVolumeSpecName: "kube-api-access-ngcfz") pod "30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" (UID: "30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0"). InnerVolumeSpecName "kube-api-access-ngcfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:10:06 crc kubenswrapper[4789]: I1008 15:10:06.962538 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" (UID: "30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.047950 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngcfz\" (UniqueName: \"kubernetes.io/projected/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-kube-api-access-ngcfz\") on node \"crc\" DevicePath \"\"" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.047982 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.048006 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.314306 4789 generic.go:334] "Generic (PLEG): container finished" podID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerID="3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6" exitCode=0 Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.314353 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8x8wr" event={"ID":"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0","Type":"ContainerDied","Data":"3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6"} Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.314385 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8x8wr" event={"ID":"30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0","Type":"ContainerDied","Data":"2758136f538c0e282f3f5f4e1b2e49b46d65b937d1ae1a89583ea7509019ef37"} Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.314390 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8x8wr" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.314407 4789 scope.go:117] "RemoveContainer" containerID="3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.337590 4789 scope.go:117] "RemoveContainer" containerID="5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.358617 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8x8wr"] Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.367347 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8x8wr"] Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.595396 4789 scope.go:117] "RemoveContainer" containerID="6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.650193 4789 scope.go:117] "RemoveContainer" containerID="3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6" Oct 08 15:10:07 crc kubenswrapper[4789]: E1008 15:10:07.650660 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6\": container with ID starting with 3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6 not found: ID does not exist" containerID="3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.650691 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6"} err="failed to get container status \"3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6\": rpc error: code = NotFound desc = could not find container \"3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6\": container with ID starting with 3e44e3946bf5ed437e6587393a70fda1114f74d887f8f12f836d753c94ea5ba6 not found: ID does not exist" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.650713 4789 scope.go:117] "RemoveContainer" containerID="5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc" Oct 08 15:10:07 crc kubenswrapper[4789]: E1008 15:10:07.651372 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc\": container with ID starting with 5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc not found: ID does not exist" containerID="5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.651395 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc"} err="failed to get container status \"5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc\": rpc error: code = NotFound desc = could not find container \"5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc\": container with ID starting with 5bd14d8cf8aa152566a748d86f6b0df109c40896c44000bf2a9d1d69147a90dc not found: ID does not exist" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.651408 4789 scope.go:117] "RemoveContainer" containerID="6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7" Oct 08 15:10:07 crc kubenswrapper[4789]: E1008 15:10:07.651731 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7\": container with ID starting with 6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7 not found: ID does not exist" containerID="6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7" Oct 08 15:10:07 crc kubenswrapper[4789]: I1008 15:10:07.651794 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7"} err="failed to get container status \"6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7\": rpc error: code = NotFound desc = could not find container \"6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7\": container with ID starting with 6b50e6ea825e653f30c458d79de6f07b2d98d65a4504f937b9cb2838653c88e7 not found: ID does not exist" Oct 08 15:10:08 crc kubenswrapper[4789]: I1008 15:10:08.745858 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" path="/var/lib/kubelet/pods/30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0/volumes" Oct 08 15:10:26 crc kubenswrapper[4789]: I1008 15:10:26.432531 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:10:26 crc kubenswrapper[4789]: I1008 15:10:26.433064 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:10:26 crc kubenswrapper[4789]: I1008 15:10:26.433105 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 15:10:26 crc kubenswrapper[4789]: I1008 15:10:26.433844 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"73a99c9d7a580b4700b76660ad13798cfcc0758d3e56372942489377a92f0c2c"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:10:26 crc kubenswrapper[4789]: I1008 15:10:26.433902 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://73a99c9d7a580b4700b76660ad13798cfcc0758d3e56372942489377a92f0c2c" gracePeriod=600 Oct 08 15:10:27 crc kubenswrapper[4789]: I1008 15:10:27.511357 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="73a99c9d7a580b4700b76660ad13798cfcc0758d3e56372942489377a92f0c2c" exitCode=0 Oct 08 15:10:27 crc kubenswrapper[4789]: I1008 15:10:27.511450 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"73a99c9d7a580b4700b76660ad13798cfcc0758d3e56372942489377a92f0c2c"} Oct 08 15:10:27 crc kubenswrapper[4789]: I1008 15:10:27.511925 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64"} Oct 08 15:10:27 crc kubenswrapper[4789]: I1008 15:10:27.511948 4789 scope.go:117] "RemoveContainer" containerID="5d78d54c616bd1487c782ff67183f62b2e195e0d19082557cbe15d3713fff1ea" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.046320 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-488rl"] Oct 08 15:12:23 crc kubenswrapper[4789]: E1008 15:12:23.047307 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerName="extract-content" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.047321 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerName="extract-content" Oct 08 15:12:23 crc kubenswrapper[4789]: E1008 15:12:23.047340 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerName="registry-server" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.047346 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerName="registry-server" Oct 08 15:12:23 crc kubenswrapper[4789]: E1008 15:12:23.047363 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerName="extract-utilities" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.047369 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerName="extract-utilities" Oct 08 15:12:23 crc kubenswrapper[4789]: E1008 15:12:23.047382 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerName="registry-server" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.047387 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerName="registry-server" Oct 08 15:12:23 crc kubenswrapper[4789]: E1008 15:12:23.047401 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerName="extract-content" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.047407 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerName="extract-content" Oct 08 15:12:23 crc kubenswrapper[4789]: E1008 15:12:23.047422 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerName="extract-utilities" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.047428 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerName="extract-utilities" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.047636 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f9b642-b86f-4815-b903-e74f1e17dc4d" containerName="registry-server" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.047657 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="30fcfaf9-6a90-4e41-9d64-6fb9b1a310a0" containerName="registry-server" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.049330 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.057855 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5276\" (UniqueName: \"kubernetes.io/projected/1d890536-1c92-4cd1-afdc-979f3bc0a3de-kube-api-access-m5276\") pod \"certified-operators-488rl\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.058104 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-catalog-content\") pod \"certified-operators-488rl\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.058357 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-utilities\") pod \"certified-operators-488rl\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.069867 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-488rl"] Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.159459 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5276\" (UniqueName: \"kubernetes.io/projected/1d890536-1c92-4cd1-afdc-979f3bc0a3de-kube-api-access-m5276\") pod \"certified-operators-488rl\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.159746 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-catalog-content\") pod \"certified-operators-488rl\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.159893 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-utilities\") pod \"certified-operators-488rl\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.160396 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-utilities\") pod \"certified-operators-488rl\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.161099 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-catalog-content\") pod \"certified-operators-488rl\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.183951 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5276\" (UniqueName: \"kubernetes.io/projected/1d890536-1c92-4cd1-afdc-979f3bc0a3de-kube-api-access-m5276\") pod \"certified-operators-488rl\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.373875 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:23 crc kubenswrapper[4789]: I1008 15:12:23.936917 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-488rl"] Oct 08 15:12:24 crc kubenswrapper[4789]: I1008 15:12:24.708834 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-488rl" event={"ID":"1d890536-1c92-4cd1-afdc-979f3bc0a3de","Type":"ContainerStarted","Data":"86380c9b90dbf5e11ceeeaa622c5090b19210efdc7a9214169d0dfca1588c5fd"} Oct 08 15:12:25 crc kubenswrapper[4789]: I1008 15:12:25.722020 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerID="849390cc734867559e40ad78230f7b46dacd8099479c1829271045d4ec950f40" exitCode=0 Oct 08 15:12:25 crc kubenswrapper[4789]: I1008 15:12:25.722090 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-488rl" event={"ID":"1d890536-1c92-4cd1-afdc-979f3bc0a3de","Type":"ContainerDied","Data":"849390cc734867559e40ad78230f7b46dacd8099479c1829271045d4ec950f40"} Oct 08 15:12:37 crc kubenswrapper[4789]: I1008 15:12:37.856213 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-488rl" event={"ID":"1d890536-1c92-4cd1-afdc-979f3bc0a3de","Type":"ContainerStarted","Data":"09921e01fa64a842fcf7647a279efa71b623272f04b1bfc72ca2f693a715cba2"} Oct 08 15:12:39 crc kubenswrapper[4789]: I1008 15:12:39.876287 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerID="09921e01fa64a842fcf7647a279efa71b623272f04b1bfc72ca2f693a715cba2" exitCode=0 Oct 08 15:12:39 crc kubenswrapper[4789]: I1008 15:12:39.876410 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-488rl" event={"ID":"1d890536-1c92-4cd1-afdc-979f3bc0a3de","Type":"ContainerDied","Data":"09921e01fa64a842fcf7647a279efa71b623272f04b1bfc72ca2f693a715cba2"} Oct 08 15:12:41 crc kubenswrapper[4789]: I1008 15:12:41.914441 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-488rl" event={"ID":"1d890536-1c92-4cd1-afdc-979f3bc0a3de","Type":"ContainerStarted","Data":"0803645f21ae9fbb190728d0dbef6bd5e614a9dbce515ab0ea3894f41e15b305"} Oct 08 15:12:41 crc kubenswrapper[4789]: I1008 15:12:41.943190 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-488rl" podStartSLOduration=3.719679895 podStartE2EDuration="18.943169722s" podCreationTimestamp="2025-10-08 15:12:23 +0000 UTC" firstStartedPulling="2025-10-08 15:12:25.725278784 +0000 UTC m=+4285.632026276" lastFinishedPulling="2025-10-08 15:12:40.948768591 +0000 UTC m=+4300.855516103" observedRunningTime="2025-10-08 15:12:41.933091097 +0000 UTC m=+4301.839838599" watchObservedRunningTime="2025-10-08 15:12:41.943169722 +0000 UTC m=+4301.849917214" Oct 08 15:12:43 crc kubenswrapper[4789]: I1008 15:12:43.375165 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:43 crc kubenswrapper[4789]: I1008 15:12:43.375528 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:43 crc kubenswrapper[4789]: I1008 15:12:43.423809 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:53 crc kubenswrapper[4789]: I1008 15:12:53.419901 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:53 crc kubenswrapper[4789]: I1008 15:12:53.465832 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-488rl"] Oct 08 15:12:54 crc kubenswrapper[4789]: I1008 15:12:54.044953 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-488rl" podUID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerName="registry-server" containerID="cri-o://0803645f21ae9fbb190728d0dbef6bd5e614a9dbce515ab0ea3894f41e15b305" gracePeriod=2 Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.055959 4789 generic.go:334] "Generic (PLEG): container finished" podID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerID="0803645f21ae9fbb190728d0dbef6bd5e614a9dbce515ab0ea3894f41e15b305" exitCode=0 Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.056076 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-488rl" event={"ID":"1d890536-1c92-4cd1-afdc-979f3bc0a3de","Type":"ContainerDied","Data":"0803645f21ae9fbb190728d0dbef6bd5e614a9dbce515ab0ea3894f41e15b305"} Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.056421 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-488rl" event={"ID":"1d890536-1c92-4cd1-afdc-979f3bc0a3de","Type":"ContainerDied","Data":"86380c9b90dbf5e11ceeeaa622c5090b19210efdc7a9214169d0dfca1588c5fd"} Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.056444 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86380c9b90dbf5e11ceeeaa622c5090b19210efdc7a9214169d0dfca1588c5fd" Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.092596 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.266398 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-utilities\") pod \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.266654 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-catalog-content\") pod \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.266730 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5276\" (UniqueName: \"kubernetes.io/projected/1d890536-1c92-4cd1-afdc-979f3bc0a3de-kube-api-access-m5276\") pod \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\" (UID: \"1d890536-1c92-4cd1-afdc-979f3bc0a3de\") " Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.267398 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-utilities" (OuterVolumeSpecName: "utilities") pod "1d890536-1c92-4cd1-afdc-979f3bc0a3de" (UID: "1d890536-1c92-4cd1-afdc-979f3bc0a3de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.279295 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d890536-1c92-4cd1-afdc-979f3bc0a3de-kube-api-access-m5276" (OuterVolumeSpecName: "kube-api-access-m5276") pod "1d890536-1c92-4cd1-afdc-979f3bc0a3de" (UID: "1d890536-1c92-4cd1-afdc-979f3bc0a3de"). InnerVolumeSpecName "kube-api-access-m5276". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.308429 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d890536-1c92-4cd1-afdc-979f3bc0a3de" (UID: "1d890536-1c92-4cd1-afdc-979f3bc0a3de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.368665 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.368694 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d890536-1c92-4cd1-afdc-979f3bc0a3de-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:12:55 crc kubenswrapper[4789]: I1008 15:12:55.368705 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5276\" (UniqueName: \"kubernetes.io/projected/1d890536-1c92-4cd1-afdc-979f3bc0a3de-kube-api-access-m5276\") on node \"crc\" DevicePath \"\"" Oct 08 15:12:56 crc kubenswrapper[4789]: I1008 15:12:56.064542 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-488rl" Oct 08 15:12:56 crc kubenswrapper[4789]: I1008 15:12:56.117061 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-488rl"] Oct 08 15:12:56 crc kubenswrapper[4789]: I1008 15:12:56.126002 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-488rl"] Oct 08 15:12:56 crc kubenswrapper[4789]: I1008 15:12:56.433036 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:12:56 crc kubenswrapper[4789]: I1008 15:12:56.433096 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:12:56 crc kubenswrapper[4789]: I1008 15:12:56.754848 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" path="/var/lib/kubelet/pods/1d890536-1c92-4cd1-afdc-979f3bc0a3de/volumes" Oct 08 15:13:26 crc kubenswrapper[4789]: I1008 15:13:26.433817 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:13:26 crc kubenswrapper[4789]: I1008 15:13:26.434394 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:13:56 crc kubenswrapper[4789]: I1008 15:13:56.433441 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:13:56 crc kubenswrapper[4789]: I1008 15:13:56.434049 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:13:56 crc kubenswrapper[4789]: I1008 15:13:56.434100 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 15:13:56 crc kubenswrapper[4789]: I1008 15:13:56.434961 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:13:56 crc kubenswrapper[4789]: I1008 15:13:56.435035 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" gracePeriod=600 Oct 08 15:13:57 crc kubenswrapper[4789]: I1008 15:13:57.679382 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" exitCode=0 Oct 08 15:13:57 crc kubenswrapper[4789]: I1008 15:13:57.679456 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64"} Oct 08 15:13:57 crc kubenswrapper[4789]: I1008 15:13:57.679716 4789 scope.go:117] "RemoveContainer" containerID="73a99c9d7a580b4700b76660ad13798cfcc0758d3e56372942489377a92f0c2c" Oct 08 15:13:57 crc kubenswrapper[4789]: E1008 15:13:57.808354 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:13:58 crc kubenswrapper[4789]: I1008 15:13:58.696857 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:13:58 crc kubenswrapper[4789]: E1008 15:13:58.697578 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:14:09 crc kubenswrapper[4789]: I1008 15:14:09.730098 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:14:09 crc kubenswrapper[4789]: E1008 15:14:09.730769 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:14:22 crc kubenswrapper[4789]: I1008 15:14:22.730068 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:14:22 crc kubenswrapper[4789]: E1008 15:14:22.730823 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:14:36 crc kubenswrapper[4789]: I1008 15:14:36.730375 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:14:36 crc kubenswrapper[4789]: E1008 15:14:36.731037 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:14:48 crc kubenswrapper[4789]: I1008 15:14:48.730200 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:14:48 crc kubenswrapper[4789]: E1008 15:14:48.730949 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.152334 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv"] Oct 08 15:15:00 crc kubenswrapper[4789]: E1008 15:15:00.154637 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerName="extract-content" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.154664 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerName="extract-content" Oct 08 15:15:00 crc kubenswrapper[4789]: E1008 15:15:00.154701 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerName="extract-utilities" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.154710 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerName="extract-utilities" Oct 08 15:15:00 crc kubenswrapper[4789]: E1008 15:15:00.154722 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerName="registry-server" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.154727 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerName="registry-server" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.155015 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d890536-1c92-4cd1-afdc-979f3bc0a3de" containerName="registry-server" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.155720 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.157869 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.159862 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.191618 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv"] Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.264395 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmbpz\" (UniqueName: \"kubernetes.io/projected/7c07b8eb-8a12-4780-a577-480d8b5b8067-kube-api-access-vmbpz\") pod \"collect-profiles-29332275-xrggv\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.264756 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c07b8eb-8a12-4780-a577-480d8b5b8067-secret-volume\") pod \"collect-profiles-29332275-xrggv\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.264883 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c07b8eb-8a12-4780-a577-480d8b5b8067-config-volume\") pod \"collect-profiles-29332275-xrggv\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.366567 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmbpz\" (UniqueName: \"kubernetes.io/projected/7c07b8eb-8a12-4780-a577-480d8b5b8067-kube-api-access-vmbpz\") pod \"collect-profiles-29332275-xrggv\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.366709 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c07b8eb-8a12-4780-a577-480d8b5b8067-secret-volume\") pod \"collect-profiles-29332275-xrggv\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.366764 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c07b8eb-8a12-4780-a577-480d8b5b8067-config-volume\") pod \"collect-profiles-29332275-xrggv\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.368910 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c07b8eb-8a12-4780-a577-480d8b5b8067-config-volume\") pod \"collect-profiles-29332275-xrggv\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.376010 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c07b8eb-8a12-4780-a577-480d8b5b8067-secret-volume\") pod \"collect-profiles-29332275-xrggv\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.382931 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmbpz\" (UniqueName: \"kubernetes.io/projected/7c07b8eb-8a12-4780-a577-480d8b5b8067-kube-api-access-vmbpz\") pod \"collect-profiles-29332275-xrggv\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:00 crc kubenswrapper[4789]: I1008 15:15:00.491234 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:01 crc kubenswrapper[4789]: I1008 15:15:01.678546 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv"] Oct 08 15:15:02 crc kubenswrapper[4789]: I1008 15:15:02.374721 4789 generic.go:334] "Generic (PLEG): container finished" podID="7c07b8eb-8a12-4780-a577-480d8b5b8067" containerID="bf639e67794d2d0030756ef11d7b2d6b7cc3f212aeb24115eb0545fafddb0bca" exitCode=0 Oct 08 15:15:02 crc kubenswrapper[4789]: I1008 15:15:02.374806 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" event={"ID":"7c07b8eb-8a12-4780-a577-480d8b5b8067","Type":"ContainerDied","Data":"bf639e67794d2d0030756ef11d7b2d6b7cc3f212aeb24115eb0545fafddb0bca"} Oct 08 15:15:02 crc kubenswrapper[4789]: I1008 15:15:02.375054 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" event={"ID":"7c07b8eb-8a12-4780-a577-480d8b5b8067","Type":"ContainerStarted","Data":"2275d90ee6e01072616def9a9d4c2d22bb0ff4c164c6b20edbbf0db6352bb6eb"} Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.730616 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:15:03 crc kubenswrapper[4789]: E1008 15:15:03.731664 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.748830 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.850866 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmbpz\" (UniqueName: \"kubernetes.io/projected/7c07b8eb-8a12-4780-a577-480d8b5b8067-kube-api-access-vmbpz\") pod \"7c07b8eb-8a12-4780-a577-480d8b5b8067\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.851192 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c07b8eb-8a12-4780-a577-480d8b5b8067-config-volume\") pod \"7c07b8eb-8a12-4780-a577-480d8b5b8067\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.851279 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c07b8eb-8a12-4780-a577-480d8b5b8067-secret-volume\") pod \"7c07b8eb-8a12-4780-a577-480d8b5b8067\" (UID: \"7c07b8eb-8a12-4780-a577-480d8b5b8067\") " Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.853186 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c07b8eb-8a12-4780-a577-480d8b5b8067-config-volume" (OuterVolumeSpecName: "config-volume") pod "7c07b8eb-8a12-4780-a577-480d8b5b8067" (UID: "7c07b8eb-8a12-4780-a577-480d8b5b8067"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.857526 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c07b8eb-8a12-4780-a577-480d8b5b8067-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7c07b8eb-8a12-4780-a577-480d8b5b8067" (UID: "7c07b8eb-8a12-4780-a577-480d8b5b8067"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.857873 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c07b8eb-8a12-4780-a577-480d8b5b8067-kube-api-access-vmbpz" (OuterVolumeSpecName: "kube-api-access-vmbpz") pod "7c07b8eb-8a12-4780-a577-480d8b5b8067" (UID: "7c07b8eb-8a12-4780-a577-480d8b5b8067"). InnerVolumeSpecName "kube-api-access-vmbpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.954777 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c07b8eb-8a12-4780-a577-480d8b5b8067-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.954812 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c07b8eb-8a12-4780-a577-480d8b5b8067-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:15:03 crc kubenswrapper[4789]: I1008 15:15:03.954823 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmbpz\" (UniqueName: \"kubernetes.io/projected/7c07b8eb-8a12-4780-a577-480d8b5b8067-kube-api-access-vmbpz\") on node \"crc\" DevicePath \"\"" Oct 08 15:15:04 crc kubenswrapper[4789]: I1008 15:15:04.393628 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" event={"ID":"7c07b8eb-8a12-4780-a577-480d8b5b8067","Type":"ContainerDied","Data":"2275d90ee6e01072616def9a9d4c2d22bb0ff4c164c6b20edbbf0db6352bb6eb"} Oct 08 15:15:04 crc kubenswrapper[4789]: I1008 15:15:04.393667 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2275d90ee6e01072616def9a9d4c2d22bb0ff4c164c6b20edbbf0db6352bb6eb" Oct 08 15:15:04 crc kubenswrapper[4789]: I1008 15:15:04.393737 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332275-xrggv" Oct 08 15:15:04 crc kubenswrapper[4789]: I1008 15:15:04.846228 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9"] Oct 08 15:15:04 crc kubenswrapper[4789]: I1008 15:15:04.854316 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332230-gx5c9"] Oct 08 15:15:06 crc kubenswrapper[4789]: I1008 15:15:06.741182 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="174b6cc4-86fa-405f-897f-a62690c145fd" path="/var/lib/kubelet/pods/174b6cc4-86fa-405f-897f-a62690c145fd/volumes" Oct 08 15:15:17 crc kubenswrapper[4789]: I1008 15:15:17.730873 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:15:17 crc kubenswrapper[4789]: E1008 15:15:17.731707 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:15:25 crc kubenswrapper[4789]: I1008 15:15:25.626337 4789 scope.go:117] "RemoveContainer" containerID="6c6a946f8166584d8b2beb3c190594b70ced2cf42ebef4357e4b1e04c12680bf" Oct 08 15:15:32 crc kubenswrapper[4789]: I1008 15:15:32.731678 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:15:32 crc kubenswrapper[4789]: E1008 15:15:32.733425 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:15:45 crc kubenswrapper[4789]: I1008 15:15:45.730383 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:15:45 crc kubenswrapper[4789]: E1008 15:15:45.731229 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:15:56 crc kubenswrapper[4789]: I1008 15:15:56.730269 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:15:56 crc kubenswrapper[4789]: E1008 15:15:56.730956 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:16:10 crc kubenswrapper[4789]: I1008 15:16:10.736383 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:16:10 crc kubenswrapper[4789]: E1008 15:16:10.737121 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.452765 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f86xd"] Oct 08 15:16:24 crc kubenswrapper[4789]: E1008 15:16:24.453898 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c07b8eb-8a12-4780-a577-480d8b5b8067" containerName="collect-profiles" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.453914 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c07b8eb-8a12-4780-a577-480d8b5b8067" containerName="collect-profiles" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.455469 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c07b8eb-8a12-4780-a577-480d8b5b8067" containerName="collect-profiles" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.457329 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.486193 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f86xd"] Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.589114 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-catalog-content\") pod \"redhat-operators-f86xd\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.589535 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-utilities\") pod \"redhat-operators-f86xd\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.589575 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgssp\" (UniqueName: \"kubernetes.io/projected/ff4c531e-6574-4d58-a06b-fe8ad53f4221-kube-api-access-wgssp\") pod \"redhat-operators-f86xd\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.691560 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-catalog-content\") pod \"redhat-operators-f86xd\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.691659 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-utilities\") pod \"redhat-operators-f86xd\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.691701 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgssp\" (UniqueName: \"kubernetes.io/projected/ff4c531e-6574-4d58-a06b-fe8ad53f4221-kube-api-access-wgssp\") pod \"redhat-operators-f86xd\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.692098 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-utilities\") pod \"redhat-operators-f86xd\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.692184 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-catalog-content\") pod \"redhat-operators-f86xd\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.717175 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgssp\" (UniqueName: \"kubernetes.io/projected/ff4c531e-6574-4d58-a06b-fe8ad53f4221-kube-api-access-wgssp\") pod \"redhat-operators-f86xd\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:24 crc kubenswrapper[4789]: I1008 15:16:24.800291 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:25 crc kubenswrapper[4789]: I1008 15:16:25.290192 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f86xd"] Oct 08 15:16:25 crc kubenswrapper[4789]: I1008 15:16:25.731228 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:16:25 crc kubenswrapper[4789]: E1008 15:16:25.731817 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:16:26 crc kubenswrapper[4789]: I1008 15:16:26.199057 4789 generic.go:334] "Generic (PLEG): container finished" podID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerID="82d29f208e707ead703f5075db199c838cd66abee07661185da9972431755f70" exitCode=0 Oct 08 15:16:26 crc kubenswrapper[4789]: I1008 15:16:26.199127 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f86xd" event={"ID":"ff4c531e-6574-4d58-a06b-fe8ad53f4221","Type":"ContainerDied","Data":"82d29f208e707ead703f5075db199c838cd66abee07661185da9972431755f70"} Oct 08 15:16:26 crc kubenswrapper[4789]: I1008 15:16:26.199410 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f86xd" event={"ID":"ff4c531e-6574-4d58-a06b-fe8ad53f4221","Type":"ContainerStarted","Data":"5b4ec659447c1717cf27a10f251597e82aa17531ce5503ce74c5b8cc315db8e3"} Oct 08 15:16:26 crc kubenswrapper[4789]: I1008 15:16:26.202829 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:16:30 crc kubenswrapper[4789]: I1008 15:16:30.240689 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f86xd" event={"ID":"ff4c531e-6574-4d58-a06b-fe8ad53f4221","Type":"ContainerStarted","Data":"8c53d16d527645d5398b2f03798074f4135a4e5fb36069574b90c7d4ce096ecc"} Oct 08 15:16:38 crc kubenswrapper[4789]: I1008 15:16:38.730213 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:16:38 crc kubenswrapper[4789]: E1008 15:16:38.731384 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:16:46 crc kubenswrapper[4789]: I1008 15:16:46.424872 4789 generic.go:334] "Generic (PLEG): container finished" podID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerID="8c53d16d527645d5398b2f03798074f4135a4e5fb36069574b90c7d4ce096ecc" exitCode=0 Oct 08 15:16:46 crc kubenswrapper[4789]: I1008 15:16:46.424976 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f86xd" event={"ID":"ff4c531e-6574-4d58-a06b-fe8ad53f4221","Type":"ContainerDied","Data":"8c53d16d527645d5398b2f03798074f4135a4e5fb36069574b90c7d4ce096ecc"} Oct 08 15:16:49 crc kubenswrapper[4789]: I1008 15:16:49.469850 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f86xd" event={"ID":"ff4c531e-6574-4d58-a06b-fe8ad53f4221","Type":"ContainerStarted","Data":"01b5fa9b02a15565e1fe021a5fc5b49f0e9d35917bfe94fbea77b9b2485cf782"} Oct 08 15:16:49 crc kubenswrapper[4789]: I1008 15:16:49.513062 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f86xd" podStartSLOduration=3.474268338 podStartE2EDuration="25.513032378s" podCreationTimestamp="2025-10-08 15:16:24 +0000 UTC" firstStartedPulling="2025-10-08 15:16:26.202390491 +0000 UTC m=+4526.109138013" lastFinishedPulling="2025-10-08 15:16:48.241154551 +0000 UTC m=+4548.147902053" observedRunningTime="2025-10-08 15:16:49.499376385 +0000 UTC m=+4549.406123927" watchObservedRunningTime="2025-10-08 15:16:49.513032378 +0000 UTC m=+4549.419779890" Oct 08 15:16:53 crc kubenswrapper[4789]: I1008 15:16:53.730442 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:16:53 crc kubenswrapper[4789]: E1008 15:16:53.731167 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:16:54 crc kubenswrapper[4789]: I1008 15:16:54.801315 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:54 crc kubenswrapper[4789]: I1008 15:16:54.801398 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:54 crc kubenswrapper[4789]: I1008 15:16:54.878822 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:55 crc kubenswrapper[4789]: I1008 15:16:55.609399 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:55 crc kubenswrapper[4789]: I1008 15:16:55.674547 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f86xd"] Oct 08 15:16:57 crc kubenswrapper[4789]: I1008 15:16:57.562413 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f86xd" podUID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerName="registry-server" containerID="cri-o://01b5fa9b02a15565e1fe021a5fc5b49f0e9d35917bfe94fbea77b9b2485cf782" gracePeriod=2 Oct 08 15:16:58 crc kubenswrapper[4789]: I1008 15:16:58.578486 4789 generic.go:334] "Generic (PLEG): container finished" podID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerID="01b5fa9b02a15565e1fe021a5fc5b49f0e9d35917bfe94fbea77b9b2485cf782" exitCode=0 Oct 08 15:16:58 crc kubenswrapper[4789]: I1008 15:16:58.578581 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f86xd" event={"ID":"ff4c531e-6574-4d58-a06b-fe8ad53f4221","Type":"ContainerDied","Data":"01b5fa9b02a15565e1fe021a5fc5b49f0e9d35917bfe94fbea77b9b2485cf782"} Oct 08 15:16:58 crc kubenswrapper[4789]: I1008 15:16:58.891807 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:58 crc kubenswrapper[4789]: I1008 15:16:58.982024 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgssp\" (UniqueName: \"kubernetes.io/projected/ff4c531e-6574-4d58-a06b-fe8ad53f4221-kube-api-access-wgssp\") pod \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " Oct 08 15:16:58 crc kubenswrapper[4789]: I1008 15:16:58.982152 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-catalog-content\") pod \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " Oct 08 15:16:58 crc kubenswrapper[4789]: I1008 15:16:58.982247 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-utilities\") pod \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\" (UID: \"ff4c531e-6574-4d58-a06b-fe8ad53f4221\") " Oct 08 15:16:58 crc kubenswrapper[4789]: I1008 15:16:58.983117 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-utilities" (OuterVolumeSpecName: "utilities") pod "ff4c531e-6574-4d58-a06b-fe8ad53f4221" (UID: "ff4c531e-6574-4d58-a06b-fe8ad53f4221"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:16:58 crc kubenswrapper[4789]: I1008 15:16:58.991161 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff4c531e-6574-4d58-a06b-fe8ad53f4221-kube-api-access-wgssp" (OuterVolumeSpecName: "kube-api-access-wgssp") pod "ff4c531e-6574-4d58-a06b-fe8ad53f4221" (UID: "ff4c531e-6574-4d58-a06b-fe8ad53f4221"). InnerVolumeSpecName "kube-api-access-wgssp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.082807 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff4c531e-6574-4d58-a06b-fe8ad53f4221" (UID: "ff4c531e-6574-4d58-a06b-fe8ad53f4221"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.084279 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.084302 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgssp\" (UniqueName: \"kubernetes.io/projected/ff4c531e-6574-4d58-a06b-fe8ad53f4221-kube-api-access-wgssp\") on node \"crc\" DevicePath \"\"" Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.084315 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff4c531e-6574-4d58-a06b-fe8ad53f4221-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.592682 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f86xd" event={"ID":"ff4c531e-6574-4d58-a06b-fe8ad53f4221","Type":"ContainerDied","Data":"5b4ec659447c1717cf27a10f251597e82aa17531ce5503ce74c5b8cc315db8e3"} Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.593266 4789 scope.go:117] "RemoveContainer" containerID="01b5fa9b02a15565e1fe021a5fc5b49f0e9d35917bfe94fbea77b9b2485cf782" Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.592796 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f86xd" Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.628554 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f86xd"] Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.638091 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f86xd"] Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.640293 4789 scope.go:117] "RemoveContainer" containerID="8c53d16d527645d5398b2f03798074f4135a4e5fb36069574b90c7d4ce096ecc" Oct 08 15:16:59 crc kubenswrapper[4789]: I1008 15:16:59.675149 4789 scope.go:117] "RemoveContainer" containerID="82d29f208e707ead703f5075db199c838cd66abee07661185da9972431755f70" Oct 08 15:17:00 crc kubenswrapper[4789]: I1008 15:17:00.744382 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" path="/var/lib/kubelet/pods/ff4c531e-6574-4d58-a06b-fe8ad53f4221/volumes" Oct 08 15:17:06 crc kubenswrapper[4789]: I1008 15:17:06.731135 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:17:06 crc kubenswrapper[4789]: E1008 15:17:06.731880 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:17:21 crc kubenswrapper[4789]: I1008 15:17:21.730563 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:17:21 crc kubenswrapper[4789]: E1008 15:17:21.731372 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:17:34 crc kubenswrapper[4789]: I1008 15:17:34.730165 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:17:34 crc kubenswrapper[4789]: E1008 15:17:34.731346 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:17:46 crc kubenswrapper[4789]: I1008 15:17:46.731088 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:17:46 crc kubenswrapper[4789]: E1008 15:17:46.732051 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:17:57 crc kubenswrapper[4789]: I1008 15:17:57.730794 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:17:57 crc kubenswrapper[4789]: E1008 15:17:57.731912 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:18:11 crc kubenswrapper[4789]: I1008 15:18:11.731092 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:18:11 crc kubenswrapper[4789]: E1008 15:18:11.733289 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:18:25 crc kubenswrapper[4789]: I1008 15:18:25.730457 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:18:25 crc kubenswrapper[4789]: E1008 15:18:25.731396 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:18:25 crc kubenswrapper[4789]: I1008 15:18:25.763243 4789 scope.go:117] "RemoveContainer" containerID="849390cc734867559e40ad78230f7b46dacd8099479c1829271045d4ec950f40" Oct 08 15:18:38 crc kubenswrapper[4789]: I1008 15:18:38.730693 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:18:38 crc kubenswrapper[4789]: E1008 15:18:38.731536 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:18:51 crc kubenswrapper[4789]: I1008 15:18:51.730272 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:18:51 crc kubenswrapper[4789]: E1008 15:18:51.731203 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:19:06 crc kubenswrapper[4789]: I1008 15:19:06.731209 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:19:07 crc kubenswrapper[4789]: I1008 15:19:07.134109 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"666e986a938fd47e63e7d617340b6cbe36f170c6d40d1953e92a7b35eaebc249"} Oct 08 15:19:25 crc kubenswrapper[4789]: I1008 15:19:25.827025 4789 scope.go:117] "RemoveContainer" containerID="09921e01fa64a842fcf7647a279efa71b623272f04b1bfc72ca2f693a715cba2" Oct 08 15:19:25 crc kubenswrapper[4789]: I1008 15:19:25.896498 4789 scope.go:117] "RemoveContainer" containerID="0803645f21ae9fbb190728d0dbef6bd5e614a9dbce515ab0ea3894f41e15b305" Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.798891 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7rtnm"] Oct 08 15:20:25 crc kubenswrapper[4789]: E1008 15:20:25.799728 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerName="extract-utilities" Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.799739 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerName="extract-utilities" Oct 08 15:20:25 crc kubenswrapper[4789]: E1008 15:20:25.799766 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerName="registry-server" Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.799772 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerName="registry-server" Oct 08 15:20:25 crc kubenswrapper[4789]: E1008 15:20:25.799792 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerName="extract-content" Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.799798 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerName="extract-content" Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.800013 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff4c531e-6574-4d58-a06b-fe8ad53f4221" containerName="registry-server" Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.801516 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.836548 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rtnm"] Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.940134 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-catalog-content\") pod \"redhat-marketplace-7rtnm\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.940232 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-utilities\") pod \"redhat-marketplace-7rtnm\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:25 crc kubenswrapper[4789]: I1008 15:20:25.940301 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2pcq\" (UniqueName: \"kubernetes.io/projected/860c56ab-570e-4788-b068-78914f9d7a80-kube-api-access-s2pcq\") pod \"redhat-marketplace-7rtnm\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.042158 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-catalog-content\") pod \"redhat-marketplace-7rtnm\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.042265 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-utilities\") pod \"redhat-marketplace-7rtnm\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.042336 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2pcq\" (UniqueName: \"kubernetes.io/projected/860c56ab-570e-4788-b068-78914f9d7a80-kube-api-access-s2pcq\") pod \"redhat-marketplace-7rtnm\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.042643 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-catalog-content\") pod \"redhat-marketplace-7rtnm\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.042735 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-utilities\") pod \"redhat-marketplace-7rtnm\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.061325 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2pcq\" (UniqueName: \"kubernetes.io/projected/860c56ab-570e-4788-b068-78914f9d7a80-kube-api-access-s2pcq\") pod \"redhat-marketplace-7rtnm\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.136969 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.661522 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rtnm"] Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.957381 4789 generic.go:334] "Generic (PLEG): container finished" podID="860c56ab-570e-4788-b068-78914f9d7a80" containerID="ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071" exitCode=0 Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.957424 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rtnm" event={"ID":"860c56ab-570e-4788-b068-78914f9d7a80","Type":"ContainerDied","Data":"ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071"} Oct 08 15:20:26 crc kubenswrapper[4789]: I1008 15:20:26.957661 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rtnm" event={"ID":"860c56ab-570e-4788-b068-78914f9d7a80","Type":"ContainerStarted","Data":"98cdc7d90355bff867c9d78ed485fdbf7c84c982ac7ec77d737112a212e3aeca"} Oct 08 15:20:28 crc kubenswrapper[4789]: I1008 15:20:28.979336 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rtnm" event={"ID":"860c56ab-570e-4788-b068-78914f9d7a80","Type":"ContainerStarted","Data":"bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63"} Oct 08 15:20:29 crc kubenswrapper[4789]: I1008 15:20:29.990053 4789 generic.go:334] "Generic (PLEG): container finished" podID="860c56ab-570e-4788-b068-78914f9d7a80" containerID="bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63" exitCode=0 Oct 08 15:20:29 crc kubenswrapper[4789]: I1008 15:20:29.990235 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rtnm" event={"ID":"860c56ab-570e-4788-b068-78914f9d7a80","Type":"ContainerDied","Data":"bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63"} Oct 08 15:20:32 crc kubenswrapper[4789]: I1008 15:20:32.013426 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rtnm" event={"ID":"860c56ab-570e-4788-b068-78914f9d7a80","Type":"ContainerStarted","Data":"7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1"} Oct 08 15:20:36 crc kubenswrapper[4789]: I1008 15:20:36.137952 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:36 crc kubenswrapper[4789]: I1008 15:20:36.138570 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:36 crc kubenswrapper[4789]: I1008 15:20:36.215025 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:36 crc kubenswrapper[4789]: I1008 15:20:36.249798 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7rtnm" podStartSLOduration=7.125060272 podStartE2EDuration="11.249781074s" podCreationTimestamp="2025-10-08 15:20:25 +0000 UTC" firstStartedPulling="2025-10-08 15:20:26.960012655 +0000 UTC m=+4766.866760147" lastFinishedPulling="2025-10-08 15:20:31.084733417 +0000 UTC m=+4770.991480949" observedRunningTime="2025-10-08 15:20:32.037293881 +0000 UTC m=+4771.944041393" watchObservedRunningTime="2025-10-08 15:20:36.249781074 +0000 UTC m=+4776.156528566" Oct 08 15:20:37 crc kubenswrapper[4789]: I1008 15:20:37.144901 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:37 crc kubenswrapper[4789]: I1008 15:20:37.209450 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rtnm"] Oct 08 15:20:39 crc kubenswrapper[4789]: I1008 15:20:39.083670 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7rtnm" podUID="860c56ab-570e-4788-b068-78914f9d7a80" containerName="registry-server" containerID="cri-o://7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1" gracePeriod=2 Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.012192 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.100304 4789 generic.go:334] "Generic (PLEG): container finished" podID="860c56ab-570e-4788-b068-78914f9d7a80" containerID="7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1" exitCode=0 Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.100362 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rtnm" event={"ID":"860c56ab-570e-4788-b068-78914f9d7a80","Type":"ContainerDied","Data":"7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1"} Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.100397 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rtnm" event={"ID":"860c56ab-570e-4788-b068-78914f9d7a80","Type":"ContainerDied","Data":"98cdc7d90355bff867c9d78ed485fdbf7c84c982ac7ec77d737112a212e3aeca"} Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.100415 4789 scope.go:117] "RemoveContainer" containerID="7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.100666 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rtnm" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.155810 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2pcq\" (UniqueName: \"kubernetes.io/projected/860c56ab-570e-4788-b068-78914f9d7a80-kube-api-access-s2pcq\") pod \"860c56ab-570e-4788-b068-78914f9d7a80\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.155931 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-utilities\") pod \"860c56ab-570e-4788-b068-78914f9d7a80\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.155973 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-catalog-content\") pod \"860c56ab-570e-4788-b068-78914f9d7a80\" (UID: \"860c56ab-570e-4788-b068-78914f9d7a80\") " Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.157292 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-utilities" (OuterVolumeSpecName: "utilities") pod "860c56ab-570e-4788-b068-78914f9d7a80" (UID: "860c56ab-570e-4788-b068-78914f9d7a80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.186583 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/860c56ab-570e-4788-b068-78914f9d7a80-kube-api-access-s2pcq" (OuterVolumeSpecName: "kube-api-access-s2pcq") pod "860c56ab-570e-4788-b068-78914f9d7a80" (UID: "860c56ab-570e-4788-b068-78914f9d7a80"). InnerVolumeSpecName "kube-api-access-s2pcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.208340 4789 scope.go:117] "RemoveContainer" containerID="bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.261460 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "860c56ab-570e-4788-b068-78914f9d7a80" (UID: "860c56ab-570e-4788-b068-78914f9d7a80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.265408 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.265430 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/860c56ab-570e-4788-b068-78914f9d7a80-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.265441 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2pcq\" (UniqueName: \"kubernetes.io/projected/860c56ab-570e-4788-b068-78914f9d7a80-kube-api-access-s2pcq\") on node \"crc\" DevicePath \"\"" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.286980 4789 scope.go:117] "RemoveContainer" containerID="ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.346753 4789 scope.go:117] "RemoveContainer" containerID="7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1" Oct 08 15:20:40 crc kubenswrapper[4789]: E1008 15:20:40.348275 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1\": container with ID starting with 7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1 not found: ID does not exist" containerID="7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.348319 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1"} err="failed to get container status \"7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1\": rpc error: code = NotFound desc = could not find container \"7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1\": container with ID starting with 7696fff316562a891ad58df5f0f7d255798669ab08e4abb5722714b56f9d16f1 not found: ID does not exist" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.348373 4789 scope.go:117] "RemoveContainer" containerID="bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63" Oct 08 15:20:40 crc kubenswrapper[4789]: E1008 15:20:40.348675 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63\": container with ID starting with bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63 not found: ID does not exist" containerID="bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.348699 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63"} err="failed to get container status \"bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63\": rpc error: code = NotFound desc = could not find container \"bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63\": container with ID starting with bd5586a2462ec7bbf1f04074ae6b6fa81788a90a58cbbc910f018ac8000aff63 not found: ID does not exist" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.348712 4789 scope.go:117] "RemoveContainer" containerID="ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071" Oct 08 15:20:40 crc kubenswrapper[4789]: E1008 15:20:40.349081 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071\": container with ID starting with ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071 not found: ID does not exist" containerID="ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.349105 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071"} err="failed to get container status \"ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071\": rpc error: code = NotFound desc = could not find container \"ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071\": container with ID starting with ebe6e9ffeb0b0177cc66ea4fbc120e06ccc2b09a7e3c0e5384dc8b26ca86f071 not found: ID does not exist" Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.435475 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rtnm"] Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.445043 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rtnm"] Oct 08 15:20:40 crc kubenswrapper[4789]: I1008 15:20:40.739580 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="860c56ab-570e-4788-b068-78914f9d7a80" path="/var/lib/kubelet/pods/860c56ab-570e-4788-b068-78914f9d7a80/volumes" Oct 08 15:21:26 crc kubenswrapper[4789]: I1008 15:21:26.433202 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:21:26 crc kubenswrapper[4789]: I1008 15:21:26.434004 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:21:56 crc kubenswrapper[4789]: I1008 15:21:56.433079 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:21:56 crc kubenswrapper[4789]: I1008 15:21:56.433767 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:22:26 crc kubenswrapper[4789]: I1008 15:22:26.432499 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:22:26 crc kubenswrapper[4789]: I1008 15:22:26.433027 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:22:26 crc kubenswrapper[4789]: I1008 15:22:26.433068 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 15:22:26 crc kubenswrapper[4789]: I1008 15:22:26.433763 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"666e986a938fd47e63e7d617340b6cbe36f170c6d40d1953e92a7b35eaebc249"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:22:26 crc kubenswrapper[4789]: I1008 15:22:26.433807 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://666e986a938fd47e63e7d617340b6cbe36f170c6d40d1953e92a7b35eaebc249" gracePeriod=600 Oct 08 15:22:27 crc kubenswrapper[4789]: I1008 15:22:27.206010 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="666e986a938fd47e63e7d617340b6cbe36f170c6d40d1953e92a7b35eaebc249" exitCode=0 Oct 08 15:22:27 crc kubenswrapper[4789]: I1008 15:22:27.206207 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"666e986a938fd47e63e7d617340b6cbe36f170c6d40d1953e92a7b35eaebc249"} Oct 08 15:22:27 crc kubenswrapper[4789]: I1008 15:22:27.206552 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4"} Oct 08 15:22:27 crc kubenswrapper[4789]: I1008 15:22:27.206574 4789 scope.go:117] "RemoveContainer" containerID="a148df465c4c8fddc9f60c50727aaac2d066cdbdcbecd566518c7fc1af409e64" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.105350 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f96jk"] Oct 08 15:22:53 crc kubenswrapper[4789]: E1008 15:22:53.106622 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860c56ab-570e-4788-b068-78914f9d7a80" containerName="extract-utilities" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.106636 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="860c56ab-570e-4788-b068-78914f9d7a80" containerName="extract-utilities" Oct 08 15:22:53 crc kubenswrapper[4789]: E1008 15:22:53.106669 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860c56ab-570e-4788-b068-78914f9d7a80" containerName="registry-server" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.106676 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="860c56ab-570e-4788-b068-78914f9d7a80" containerName="registry-server" Oct 08 15:22:53 crc kubenswrapper[4789]: E1008 15:22:53.106703 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860c56ab-570e-4788-b068-78914f9d7a80" containerName="extract-content" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.106709 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="860c56ab-570e-4788-b068-78914f9d7a80" containerName="extract-content" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.106940 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="860c56ab-570e-4788-b068-78914f9d7a80" containerName="registry-server" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.109201 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.131216 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f96jk"] Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.214853 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bdc52e6-9e72-411a-9558-0607f78fe55c-catalog-content\") pod \"certified-operators-f96jk\" (UID: \"6bdc52e6-9e72-411a-9558-0607f78fe55c\") " pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.214976 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bdc52e6-9e72-411a-9558-0607f78fe55c-utilities\") pod \"certified-operators-f96jk\" (UID: \"6bdc52e6-9e72-411a-9558-0607f78fe55c\") " pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.215045 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4khh\" (UniqueName: \"kubernetes.io/projected/6bdc52e6-9e72-411a-9558-0607f78fe55c-kube-api-access-t4khh\") pod \"certified-operators-f96jk\" (UID: \"6bdc52e6-9e72-411a-9558-0607f78fe55c\") " pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.304663 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qztcr"] Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.306844 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.318075 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bdc52e6-9e72-411a-9558-0607f78fe55c-catalog-content\") pod \"certified-operators-f96jk\" (UID: \"6bdc52e6-9e72-411a-9558-0607f78fe55c\") " pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.318192 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bdc52e6-9e72-411a-9558-0607f78fe55c-utilities\") pod \"certified-operators-f96jk\" (UID: \"6bdc52e6-9e72-411a-9558-0607f78fe55c\") " pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.318240 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4khh\" (UniqueName: \"kubernetes.io/projected/6bdc52e6-9e72-411a-9558-0607f78fe55c-kube-api-access-t4khh\") pod \"certified-operators-f96jk\" (UID: \"6bdc52e6-9e72-411a-9558-0607f78fe55c\") " pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.319375 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bdc52e6-9e72-411a-9558-0607f78fe55c-catalog-content\") pod \"certified-operators-f96jk\" (UID: \"6bdc52e6-9e72-411a-9558-0607f78fe55c\") " pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.319545 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bdc52e6-9e72-411a-9558-0607f78fe55c-utilities\") pod \"certified-operators-f96jk\" (UID: \"6bdc52e6-9e72-411a-9558-0607f78fe55c\") " pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.330526 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qztcr"] Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.372560 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4khh\" (UniqueName: \"kubernetes.io/projected/6bdc52e6-9e72-411a-9558-0607f78fe55c-kube-api-access-t4khh\") pod \"certified-operators-f96jk\" (UID: \"6bdc52e6-9e72-411a-9558-0607f78fe55c\") " pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.421324 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/110bf958-13b6-4727-9d9b-ff08e0ac9146-utilities\") pod \"community-operators-qztcr\" (UID: \"110bf958-13b6-4727-9d9b-ff08e0ac9146\") " pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.421418 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/110bf958-13b6-4727-9d9b-ff08e0ac9146-catalog-content\") pod \"community-operators-qztcr\" (UID: \"110bf958-13b6-4727-9d9b-ff08e0ac9146\") " pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.421500 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v77cz\" (UniqueName: \"kubernetes.io/projected/110bf958-13b6-4727-9d9b-ff08e0ac9146-kube-api-access-v77cz\") pod \"community-operators-qztcr\" (UID: \"110bf958-13b6-4727-9d9b-ff08e0ac9146\") " pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.450785 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.523395 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/110bf958-13b6-4727-9d9b-ff08e0ac9146-utilities\") pod \"community-operators-qztcr\" (UID: \"110bf958-13b6-4727-9d9b-ff08e0ac9146\") " pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.523811 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/110bf958-13b6-4727-9d9b-ff08e0ac9146-catalog-content\") pod \"community-operators-qztcr\" (UID: \"110bf958-13b6-4727-9d9b-ff08e0ac9146\") " pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.523901 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v77cz\" (UniqueName: \"kubernetes.io/projected/110bf958-13b6-4727-9d9b-ff08e0ac9146-kube-api-access-v77cz\") pod \"community-operators-qztcr\" (UID: \"110bf958-13b6-4727-9d9b-ff08e0ac9146\") " pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.524209 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/110bf958-13b6-4727-9d9b-ff08e0ac9146-utilities\") pod \"community-operators-qztcr\" (UID: \"110bf958-13b6-4727-9d9b-ff08e0ac9146\") " pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.524518 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/110bf958-13b6-4727-9d9b-ff08e0ac9146-catalog-content\") pod \"community-operators-qztcr\" (UID: \"110bf958-13b6-4727-9d9b-ff08e0ac9146\") " pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.550763 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v77cz\" (UniqueName: \"kubernetes.io/projected/110bf958-13b6-4727-9d9b-ff08e0ac9146-kube-api-access-v77cz\") pod \"community-operators-qztcr\" (UID: \"110bf958-13b6-4727-9d9b-ff08e0ac9146\") " pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:53 crc kubenswrapper[4789]: I1008 15:22:53.630138 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:22:54 crc kubenswrapper[4789]: I1008 15:22:54.153674 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f96jk"] Oct 08 15:22:54 crc kubenswrapper[4789]: I1008 15:22:54.318739 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qztcr"] Oct 08 15:22:54 crc kubenswrapper[4789]: W1008 15:22:54.326733 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod110bf958_13b6_4727_9d9b_ff08e0ac9146.slice/crio-3d1c73ba0156741ddefcccdc243b0b8bbfc4f7b27ad27d92f51605969a1be1ba WatchSource:0}: Error finding container 3d1c73ba0156741ddefcccdc243b0b8bbfc4f7b27ad27d92f51605969a1be1ba: Status 404 returned error can't find the container with id 3d1c73ba0156741ddefcccdc243b0b8bbfc4f7b27ad27d92f51605969a1be1ba Oct 08 15:22:54 crc kubenswrapper[4789]: I1008 15:22:54.549733 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qztcr" event={"ID":"110bf958-13b6-4727-9d9b-ff08e0ac9146","Type":"ContainerStarted","Data":"3d1c73ba0156741ddefcccdc243b0b8bbfc4f7b27ad27d92f51605969a1be1ba"} Oct 08 15:22:54 crc kubenswrapper[4789]: I1008 15:22:54.551511 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f96jk" event={"ID":"6bdc52e6-9e72-411a-9558-0607f78fe55c","Type":"ContainerStarted","Data":"a5ad3a2a003a12f906a8875469ea0a0c8049beb77a98f2a31db62cc8b7104eac"} Oct 08 15:22:55 crc kubenswrapper[4789]: I1008 15:22:55.564487 4789 generic.go:334] "Generic (PLEG): container finished" podID="6bdc52e6-9e72-411a-9558-0607f78fe55c" containerID="d098328b4e665674a3bd9d0abd9c0f641733d1cf8af44aeceb0e9fcdfb6ee911" exitCode=0 Oct 08 15:22:55 crc kubenswrapper[4789]: I1008 15:22:55.564578 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f96jk" event={"ID":"6bdc52e6-9e72-411a-9558-0607f78fe55c","Type":"ContainerDied","Data":"d098328b4e665674a3bd9d0abd9c0f641733d1cf8af44aeceb0e9fcdfb6ee911"} Oct 08 15:22:55 crc kubenswrapper[4789]: I1008 15:22:55.567274 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:22:55 crc kubenswrapper[4789]: I1008 15:22:55.568927 4789 generic.go:334] "Generic (PLEG): container finished" podID="110bf958-13b6-4727-9d9b-ff08e0ac9146" containerID="54ca09ab9233f8d79c647f4a62b765394f51bb4942c2e37d7702e638c406c3c3" exitCode=0 Oct 08 15:22:55 crc kubenswrapper[4789]: I1008 15:22:55.569018 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qztcr" event={"ID":"110bf958-13b6-4727-9d9b-ff08e0ac9146","Type":"ContainerDied","Data":"54ca09ab9233f8d79c647f4a62b765394f51bb4942c2e37d7702e638c406c3c3"} Oct 08 15:23:10 crc kubenswrapper[4789]: E1008 15:23:10.872557 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 15:23:10 crc kubenswrapper[4789]: E1008 15:23:10.873342 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t4khh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-f96jk_openshift-marketplace(6bdc52e6-9e72-411a-9558-0607f78fe55c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 15:23:10 crc kubenswrapper[4789]: E1008 15:23:10.876454 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-f96jk" podUID="6bdc52e6-9e72-411a-9558-0607f78fe55c" Oct 08 15:23:14 crc kubenswrapper[4789]: E1008 15:23:14.227251 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-f96jk" podUID="6bdc52e6-9e72-411a-9558-0607f78fe55c" Oct 08 15:23:14 crc kubenswrapper[4789]: E1008 15:23:14.298851 4789 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 15:23:14 crc kubenswrapper[4789]: E1008 15:23:14.299008 4789 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v77cz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-qztcr_openshift-marketplace(110bf958-13b6-4727-9d9b-ff08e0ac9146): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 15:23:14 crc kubenswrapper[4789]: E1008 15:23:14.300195 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-qztcr" podUID="110bf958-13b6-4727-9d9b-ff08e0ac9146" Oct 08 15:23:14 crc kubenswrapper[4789]: E1008 15:23:14.769357 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-qztcr" podUID="110bf958-13b6-4727-9d9b-ff08e0ac9146" Oct 08 15:23:34 crc kubenswrapper[4789]: I1008 15:23:34.975697 4789 generic.go:334] "Generic (PLEG): container finished" podID="110bf958-13b6-4727-9d9b-ff08e0ac9146" containerID="dbe825779d6f91e3a32f1ae0f1bbdb09aee288ae8f7c6d244127ea13fc9f4986" exitCode=0 Oct 08 15:23:34 crc kubenswrapper[4789]: I1008 15:23:34.975787 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qztcr" event={"ID":"110bf958-13b6-4727-9d9b-ff08e0ac9146","Type":"ContainerDied","Data":"dbe825779d6f91e3a32f1ae0f1bbdb09aee288ae8f7c6d244127ea13fc9f4986"} Oct 08 15:23:34 crc kubenswrapper[4789]: I1008 15:23:34.979434 4789 generic.go:334] "Generic (PLEG): container finished" podID="6bdc52e6-9e72-411a-9558-0607f78fe55c" containerID="def91242ffb5120b2ac51210fbece2a28da0ff008ff1aa2fd59746e6a1ce28e2" exitCode=0 Oct 08 15:23:34 crc kubenswrapper[4789]: I1008 15:23:34.979501 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f96jk" event={"ID":"6bdc52e6-9e72-411a-9558-0607f78fe55c","Type":"ContainerDied","Data":"def91242ffb5120b2ac51210fbece2a28da0ff008ff1aa2fd59746e6a1ce28e2"} Oct 08 15:23:45 crc kubenswrapper[4789]: I1008 15:23:45.076705 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f96jk" event={"ID":"6bdc52e6-9e72-411a-9558-0607f78fe55c","Type":"ContainerStarted","Data":"e1e00f4e25ee4d986ea591ae171539068e043d76402abca51fb688f8fcb00529"} Oct 08 15:23:45 crc kubenswrapper[4789]: I1008 15:23:45.081758 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qztcr" event={"ID":"110bf958-13b6-4727-9d9b-ff08e0ac9146","Type":"ContainerStarted","Data":"45543caabe03833c99119c3fd3a9d9680651d872c64b5e84ad71d415c5bfb97f"} Oct 08 15:23:45 crc kubenswrapper[4789]: I1008 15:23:45.099692 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f96jk" podStartSLOduration=3.962757554 podStartE2EDuration="52.099673265s" podCreationTimestamp="2025-10-08 15:22:53 +0000 UTC" firstStartedPulling="2025-10-08 15:22:55.567053854 +0000 UTC m=+4915.473801346" lastFinishedPulling="2025-10-08 15:23:43.703969565 +0000 UTC m=+4963.610717057" observedRunningTime="2025-10-08 15:23:45.093357393 +0000 UTC m=+4965.000104885" watchObservedRunningTime="2025-10-08 15:23:45.099673265 +0000 UTC m=+4965.006420747" Oct 08 15:23:45 crc kubenswrapper[4789]: I1008 15:23:45.117782 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qztcr" podStartSLOduration=3.984047133 podStartE2EDuration="52.117758457s" podCreationTimestamp="2025-10-08 15:22:53 +0000 UTC" firstStartedPulling="2025-10-08 15:22:55.572397099 +0000 UTC m=+4915.479144591" lastFinishedPulling="2025-10-08 15:23:43.706108423 +0000 UTC m=+4963.612855915" observedRunningTime="2025-10-08 15:23:45.112282258 +0000 UTC m=+4965.019029770" watchObservedRunningTime="2025-10-08 15:23:45.117758457 +0000 UTC m=+4965.024505949" Oct 08 15:23:53 crc kubenswrapper[4789]: I1008 15:23:53.451427 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:23:53 crc kubenswrapper[4789]: I1008 15:23:53.452026 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:23:53 crc kubenswrapper[4789]: I1008 15:23:53.498493 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:23:53 crc kubenswrapper[4789]: I1008 15:23:53.630846 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:23:53 crc kubenswrapper[4789]: I1008 15:23:53.630884 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:23:54 crc kubenswrapper[4789]: I1008 15:23:54.265224 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f96jk" Oct 08 15:23:54 crc kubenswrapper[4789]: I1008 15:23:54.675714 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-qztcr" podUID="110bf958-13b6-4727-9d9b-ff08e0ac9146" containerName="registry-server" probeResult="failure" output=< Oct 08 15:23:54 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 15:23:54 crc kubenswrapper[4789]: > Oct 08 15:23:55 crc kubenswrapper[4789]: I1008 15:23:55.165227 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f96jk"] Oct 08 15:23:55 crc kubenswrapper[4789]: I1008 15:23:55.324274 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dlsnl"] Oct 08 15:23:55 crc kubenswrapper[4789]: I1008 15:23:55.324622 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dlsnl" podUID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerName="registry-server" containerID="cri-o://ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89" gracePeriod=2 Oct 08 15:23:55 crc kubenswrapper[4789]: I1008 15:23:55.990489 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.078879 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n44wz\" (UniqueName: \"kubernetes.io/projected/1695b8b4-119e-4626-8e40-5ecd20d08259-kube-api-access-n44wz\") pod \"1695b8b4-119e-4626-8e40-5ecd20d08259\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.079107 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-utilities\") pod \"1695b8b4-119e-4626-8e40-5ecd20d08259\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.079144 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-catalog-content\") pod \"1695b8b4-119e-4626-8e40-5ecd20d08259\" (UID: \"1695b8b4-119e-4626-8e40-5ecd20d08259\") " Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.081123 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-utilities" (OuterVolumeSpecName: "utilities") pod "1695b8b4-119e-4626-8e40-5ecd20d08259" (UID: "1695b8b4-119e-4626-8e40-5ecd20d08259"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.106099 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1695b8b4-119e-4626-8e40-5ecd20d08259-kube-api-access-n44wz" (OuterVolumeSpecName: "kube-api-access-n44wz") pod "1695b8b4-119e-4626-8e40-5ecd20d08259" (UID: "1695b8b4-119e-4626-8e40-5ecd20d08259"). InnerVolumeSpecName "kube-api-access-n44wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.181442 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.181479 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n44wz\" (UniqueName: \"kubernetes.io/projected/1695b8b4-119e-4626-8e40-5ecd20d08259-kube-api-access-n44wz\") on node \"crc\" DevicePath \"\"" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.202361 4789 generic.go:334] "Generic (PLEG): container finished" podID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerID="ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89" exitCode=0 Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.202398 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlsnl" event={"ID":"1695b8b4-119e-4626-8e40-5ecd20d08259","Type":"ContainerDied","Data":"ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89"} Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.202445 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlsnl" event={"ID":"1695b8b4-119e-4626-8e40-5ecd20d08259","Type":"ContainerDied","Data":"8d85482fd83b89e47357f279edae0a9b77cf10abd681b2ec328b588e30fa8d40"} Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.202458 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlsnl" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.202462 4789 scope.go:117] "RemoveContainer" containerID="ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.209779 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1695b8b4-119e-4626-8e40-5ecd20d08259" (UID: "1695b8b4-119e-4626-8e40-5ecd20d08259"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.238498 4789 scope.go:117] "RemoveContainer" containerID="b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.270363 4789 scope.go:117] "RemoveContainer" containerID="58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.283757 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1695b8b4-119e-4626-8e40-5ecd20d08259-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.324782 4789 scope.go:117] "RemoveContainer" containerID="ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89" Oct 08 15:23:56 crc kubenswrapper[4789]: E1008 15:23:56.325338 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89\": container with ID starting with ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89 not found: ID does not exist" containerID="ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.325379 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89"} err="failed to get container status \"ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89\": rpc error: code = NotFound desc = could not find container \"ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89\": container with ID starting with ebd22bcdad4ba6ad7bd229c9b969bd006376c6714180a6d1aa29fcc4669cde89 not found: ID does not exist" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.325400 4789 scope.go:117] "RemoveContainer" containerID="b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560" Oct 08 15:23:56 crc kubenswrapper[4789]: E1008 15:23:56.327510 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560\": container with ID starting with b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560 not found: ID does not exist" containerID="b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.327550 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560"} err="failed to get container status \"b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560\": rpc error: code = NotFound desc = could not find container \"b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560\": container with ID starting with b5473bf4a2299ccb43b92210a0be3611713bbb9abf5dd814f1b0794ace2ad560 not found: ID does not exist" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.327572 4789 scope.go:117] "RemoveContainer" containerID="58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a" Oct 08 15:23:56 crc kubenswrapper[4789]: E1008 15:23:56.327816 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a\": container with ID starting with 58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a not found: ID does not exist" containerID="58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.327847 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a"} err="failed to get container status \"58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a\": rpc error: code = NotFound desc = could not find container \"58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a\": container with ID starting with 58285bfb822b18eb5189020be93cccbedd425f3332a953e0171a41006f68bc3a not found: ID does not exist" Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.537865 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dlsnl"] Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.545968 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dlsnl"] Oct 08 15:23:56 crc kubenswrapper[4789]: I1008 15:23:56.741999 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1695b8b4-119e-4626-8e40-5ecd20d08259" path="/var/lib/kubelet/pods/1695b8b4-119e-4626-8e40-5ecd20d08259/volumes" Oct 08 15:24:04 crc kubenswrapper[4789]: I1008 15:24:04.676720 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-qztcr" podUID="110bf958-13b6-4727-9d9b-ff08e0ac9146" containerName="registry-server" probeResult="failure" output=< Oct 08 15:24:04 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 15:24:04 crc kubenswrapper[4789]: > Oct 08 15:24:13 crc kubenswrapper[4789]: I1008 15:24:13.676204 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:24:13 crc kubenswrapper[4789]: I1008 15:24:13.733860 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qztcr" Oct 08 15:24:14 crc kubenswrapper[4789]: I1008 15:24:14.756122 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qztcr"] Oct 08 15:24:14 crc kubenswrapper[4789]: I1008 15:24:14.922467 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-flhdr"] Oct 08 15:24:14 crc kubenswrapper[4789]: I1008 15:24:14.922852 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-flhdr" podUID="673f91dd-cda6-4e43-813f-644ac69de215" containerName="registry-server" containerID="cri-o://ee44700090d88c7d7084272dd9d8e650b609188d60e5dcc2fb28fc71cf86aef8" gracePeriod=2 Oct 08 15:24:15 crc kubenswrapper[4789]: I1008 15:24:15.393610 4789 generic.go:334] "Generic (PLEG): container finished" podID="673f91dd-cda6-4e43-813f-644ac69de215" containerID="ee44700090d88c7d7084272dd9d8e650b609188d60e5dcc2fb28fc71cf86aef8" exitCode=0 Oct 08 15:24:15 crc kubenswrapper[4789]: I1008 15:24:15.394054 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flhdr" event={"ID":"673f91dd-cda6-4e43-813f-644ac69de215","Type":"ContainerDied","Data":"ee44700090d88c7d7084272dd9d8e650b609188d60e5dcc2fb28fc71cf86aef8"} Oct 08 15:24:15 crc kubenswrapper[4789]: I1008 15:24:15.874593 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flhdr" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.005413 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g55b\" (UniqueName: \"kubernetes.io/projected/673f91dd-cda6-4e43-813f-644ac69de215-kube-api-access-7g55b\") pod \"673f91dd-cda6-4e43-813f-644ac69de215\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.005575 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-catalog-content\") pod \"673f91dd-cda6-4e43-813f-644ac69de215\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.005598 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-utilities\") pod \"673f91dd-cda6-4e43-813f-644ac69de215\" (UID: \"673f91dd-cda6-4e43-813f-644ac69de215\") " Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.006195 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-utilities" (OuterVolumeSpecName: "utilities") pod "673f91dd-cda6-4e43-813f-644ac69de215" (UID: "673f91dd-cda6-4e43-813f-644ac69de215"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.022360 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/673f91dd-cda6-4e43-813f-644ac69de215-kube-api-access-7g55b" (OuterVolumeSpecName: "kube-api-access-7g55b") pod "673f91dd-cda6-4e43-813f-644ac69de215" (UID: "673f91dd-cda6-4e43-813f-644ac69de215"). InnerVolumeSpecName "kube-api-access-7g55b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.076040 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "673f91dd-cda6-4e43-813f-644ac69de215" (UID: "673f91dd-cda6-4e43-813f-644ac69de215"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.108489 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g55b\" (UniqueName: \"kubernetes.io/projected/673f91dd-cda6-4e43-813f-644ac69de215-kube-api-access-7g55b\") on node \"crc\" DevicePath \"\"" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.108521 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.108531 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/673f91dd-cda6-4e43-813f-644ac69de215-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.404470 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-flhdr" event={"ID":"673f91dd-cda6-4e43-813f-644ac69de215","Type":"ContainerDied","Data":"6ce5358441185d3f7eb02b31e193d47c1d05346aed716cf1982f79eedcf9b9eb"} Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.404506 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-flhdr" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.404536 4789 scope.go:117] "RemoveContainer" containerID="ee44700090d88c7d7084272dd9d8e650b609188d60e5dcc2fb28fc71cf86aef8" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.427760 4789 scope.go:117] "RemoveContainer" containerID="b9ec20d356e069830cb0a43fcf1293d17a0d80d4490088f396ceefbe446ef6af" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.436616 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-flhdr"] Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.447549 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-flhdr"] Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.467928 4789 scope.go:117] "RemoveContainer" containerID="bf6341b95af736f0493a4a252079c4d39846a599165c41996c7b004273c39b14" Oct 08 15:24:16 crc kubenswrapper[4789]: I1008 15:24:16.740023 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="673f91dd-cda6-4e43-813f-644ac69de215" path="/var/lib/kubelet/pods/673f91dd-cda6-4e43-813f-644ac69de215/volumes" Oct 08 15:24:26 crc kubenswrapper[4789]: I1008 15:24:26.432822 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:24:26 crc kubenswrapper[4789]: I1008 15:24:26.433570 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:24:56 crc kubenswrapper[4789]: I1008 15:24:56.433569 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:24:56 crc kubenswrapper[4789]: I1008 15:24:56.435949 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:25:26 crc kubenswrapper[4789]: I1008 15:25:26.432966 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:25:26 crc kubenswrapper[4789]: I1008 15:25:26.433833 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:25:26 crc kubenswrapper[4789]: I1008 15:25:26.433894 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 15:25:26 crc kubenswrapper[4789]: I1008 15:25:26.434710 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:25:26 crc kubenswrapper[4789]: I1008 15:25:26.434765 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" gracePeriod=600 Oct 08 15:25:27 crc kubenswrapper[4789]: I1008 15:25:27.073063 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" exitCode=0 Oct 08 15:25:27 crc kubenswrapper[4789]: I1008 15:25:27.073095 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4"} Oct 08 15:25:27 crc kubenswrapper[4789]: I1008 15:25:27.073399 4789 scope.go:117] "RemoveContainer" containerID="666e986a938fd47e63e7d617340b6cbe36f170c6d40d1953e92a7b35eaebc249" Oct 08 15:25:27 crc kubenswrapper[4789]: E1008 15:25:27.133703 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:25:28 crc kubenswrapper[4789]: I1008 15:25:28.097951 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:25:28 crc kubenswrapper[4789]: E1008 15:25:28.098264 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:25:39 crc kubenswrapper[4789]: I1008 15:25:39.730622 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:25:39 crc kubenswrapper[4789]: E1008 15:25:39.731730 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:25:52 crc kubenswrapper[4789]: I1008 15:25:52.730012 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:25:52 crc kubenswrapper[4789]: E1008 15:25:52.730841 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:26:06 crc kubenswrapper[4789]: I1008 15:26:06.731196 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:26:06 crc kubenswrapper[4789]: E1008 15:26:06.732104 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:26:18 crc kubenswrapper[4789]: I1008 15:26:18.730177 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:26:18 crc kubenswrapper[4789]: E1008 15:26:18.731034 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:26:32 crc kubenswrapper[4789]: I1008 15:26:32.730197 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:26:32 crc kubenswrapper[4789]: E1008 15:26:32.731096 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:26:43 crc kubenswrapper[4789]: I1008 15:26:43.730377 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:26:43 crc kubenswrapper[4789]: E1008 15:26:43.733377 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:26:54 crc kubenswrapper[4789]: I1008 15:26:54.730648 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:26:54 crc kubenswrapper[4789]: E1008 15:26:54.732021 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:27:05 crc kubenswrapper[4789]: I1008 15:27:05.730320 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:27:05 crc kubenswrapper[4789]: E1008 15:27:05.731164 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:27:18 crc kubenswrapper[4789]: I1008 15:27:18.730612 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:27:18 crc kubenswrapper[4789]: E1008 15:27:18.731400 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:27:29 crc kubenswrapper[4789]: I1008 15:27:29.730411 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:27:29 crc kubenswrapper[4789]: E1008 15:27:29.731320 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.449967 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-srnl8"] Oct 08 15:27:35 crc kubenswrapper[4789]: E1008 15:27:35.451188 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerName="registry-server" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.451209 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerName="registry-server" Oct 08 15:27:35 crc kubenswrapper[4789]: E1008 15:27:35.451255 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673f91dd-cda6-4e43-813f-644ac69de215" containerName="extract-content" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.451262 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="673f91dd-cda6-4e43-813f-644ac69de215" containerName="extract-content" Oct 08 15:27:35 crc kubenswrapper[4789]: E1008 15:27:35.451284 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerName="extract-utilities" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.451292 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerName="extract-utilities" Oct 08 15:27:35 crc kubenswrapper[4789]: E1008 15:27:35.451308 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673f91dd-cda6-4e43-813f-644ac69de215" containerName="extract-utilities" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.451317 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="673f91dd-cda6-4e43-813f-644ac69de215" containerName="extract-utilities" Oct 08 15:27:35 crc kubenswrapper[4789]: E1008 15:27:35.451331 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerName="extract-content" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.451338 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerName="extract-content" Oct 08 15:27:35 crc kubenswrapper[4789]: E1008 15:27:35.451361 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673f91dd-cda6-4e43-813f-644ac69de215" containerName="registry-server" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.451369 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="673f91dd-cda6-4e43-813f-644ac69de215" containerName="registry-server" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.451603 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="1695b8b4-119e-4626-8e40-5ecd20d08259" containerName="registry-server" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.451620 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="673f91dd-cda6-4e43-813f-644ac69de215" containerName="registry-server" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.460582 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.467531 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-srnl8"] Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.578931 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6ptg\" (UniqueName: \"kubernetes.io/projected/764b0312-c3a0-4569-bea1-f076aee6e61f-kube-api-access-w6ptg\") pod \"redhat-operators-srnl8\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.579543 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-catalog-content\") pod \"redhat-operators-srnl8\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.580312 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-utilities\") pod \"redhat-operators-srnl8\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.683124 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-catalog-content\") pod \"redhat-operators-srnl8\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.683296 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-utilities\") pod \"redhat-operators-srnl8\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.683404 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6ptg\" (UniqueName: \"kubernetes.io/projected/764b0312-c3a0-4569-bea1-f076aee6e61f-kube-api-access-w6ptg\") pod \"redhat-operators-srnl8\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.684151 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-catalog-content\") pod \"redhat-operators-srnl8\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.684277 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-utilities\") pod \"redhat-operators-srnl8\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.708572 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6ptg\" (UniqueName: \"kubernetes.io/projected/764b0312-c3a0-4569-bea1-f076aee6e61f-kube-api-access-w6ptg\") pod \"redhat-operators-srnl8\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:35 crc kubenswrapper[4789]: I1008 15:27:35.782964 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:36 crc kubenswrapper[4789]: I1008 15:27:36.267405 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-srnl8"] Oct 08 15:27:36 crc kubenswrapper[4789]: I1008 15:27:36.321015 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-srnl8" event={"ID":"764b0312-c3a0-4569-bea1-f076aee6e61f","Type":"ContainerStarted","Data":"1c1860df98193471b88bdb10028724d4be39e0d54c95bea1f46a93a3840b4fae"} Oct 08 15:27:37 crc kubenswrapper[4789]: I1008 15:27:37.333479 4789 generic.go:334] "Generic (PLEG): container finished" podID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerID="4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d" exitCode=0 Oct 08 15:27:37 crc kubenswrapper[4789]: I1008 15:27:37.333637 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-srnl8" event={"ID":"764b0312-c3a0-4569-bea1-f076aee6e61f","Type":"ContainerDied","Data":"4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d"} Oct 08 15:27:38 crc kubenswrapper[4789]: I1008 15:27:38.345119 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-srnl8" event={"ID":"764b0312-c3a0-4569-bea1-f076aee6e61f","Type":"ContainerStarted","Data":"00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429"} Oct 08 15:27:44 crc kubenswrapper[4789]: I1008 15:27:44.730712 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:27:44 crc kubenswrapper[4789]: E1008 15:27:44.731570 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:27:45 crc kubenswrapper[4789]: I1008 15:27:45.410422 4789 generic.go:334] "Generic (PLEG): container finished" podID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerID="00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429" exitCode=0 Oct 08 15:27:45 crc kubenswrapper[4789]: I1008 15:27:45.410515 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-srnl8" event={"ID":"764b0312-c3a0-4569-bea1-f076aee6e61f","Type":"ContainerDied","Data":"00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429"} Oct 08 15:27:46 crc kubenswrapper[4789]: I1008 15:27:46.426790 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-srnl8" event={"ID":"764b0312-c3a0-4569-bea1-f076aee6e61f","Type":"ContainerStarted","Data":"47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b"} Oct 08 15:27:55 crc kubenswrapper[4789]: I1008 15:27:55.783117 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:55 crc kubenswrapper[4789]: I1008 15:27:55.783549 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:27:56 crc kubenswrapper[4789]: I1008 15:27:56.842504 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-srnl8" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="registry-server" probeResult="failure" output=< Oct 08 15:27:56 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 15:27:56 crc kubenswrapper[4789]: > Oct 08 15:27:59 crc kubenswrapper[4789]: I1008 15:27:59.730049 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:27:59 crc kubenswrapper[4789]: E1008 15:27:59.730528 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:28:06 crc kubenswrapper[4789]: I1008 15:28:06.828122 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-srnl8" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="registry-server" probeResult="failure" output=< Oct 08 15:28:06 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 15:28:06 crc kubenswrapper[4789]: > Oct 08 15:28:14 crc kubenswrapper[4789]: I1008 15:28:14.730108 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:28:14 crc kubenswrapper[4789]: E1008 15:28:14.730889 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:28:15 crc kubenswrapper[4789]: I1008 15:28:15.835000 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:28:15 crc kubenswrapper[4789]: I1008 15:28:15.856594 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-srnl8" podStartSLOduration=32.143658823 podStartE2EDuration="40.856572486s" podCreationTimestamp="2025-10-08 15:27:35 +0000 UTC" firstStartedPulling="2025-10-08 15:27:37.335606518 +0000 UTC m=+5197.242354010" lastFinishedPulling="2025-10-08 15:27:46.048520151 +0000 UTC m=+5205.955267673" observedRunningTime="2025-10-08 15:27:46.449386745 +0000 UTC m=+5206.356134237" watchObservedRunningTime="2025-10-08 15:28:15.856572486 +0000 UTC m=+5235.763319978" Oct 08 15:28:15 crc kubenswrapper[4789]: I1008 15:28:15.879419 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:28:16 crc kubenswrapper[4789]: I1008 15:28:16.080830 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-srnl8"] Oct 08 15:28:17 crc kubenswrapper[4789]: I1008 15:28:17.740901 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-srnl8" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="registry-server" containerID="cri-o://47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b" gracePeriod=2 Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.264558 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.360913 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6ptg\" (UniqueName: \"kubernetes.io/projected/764b0312-c3a0-4569-bea1-f076aee6e61f-kube-api-access-w6ptg\") pod \"764b0312-c3a0-4569-bea1-f076aee6e61f\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.362293 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-utilities\") pod \"764b0312-c3a0-4569-bea1-f076aee6e61f\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.362367 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-catalog-content\") pod \"764b0312-c3a0-4569-bea1-f076aee6e61f\" (UID: \"764b0312-c3a0-4569-bea1-f076aee6e61f\") " Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.364900 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-utilities" (OuterVolumeSpecName: "utilities") pod "764b0312-c3a0-4569-bea1-f076aee6e61f" (UID: "764b0312-c3a0-4569-bea1-f076aee6e61f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.371794 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/764b0312-c3a0-4569-bea1-f076aee6e61f-kube-api-access-w6ptg" (OuterVolumeSpecName: "kube-api-access-w6ptg") pod "764b0312-c3a0-4569-bea1-f076aee6e61f" (UID: "764b0312-c3a0-4569-bea1-f076aee6e61f"). InnerVolumeSpecName "kube-api-access-w6ptg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.446771 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "764b0312-c3a0-4569-bea1-f076aee6e61f" (UID: "764b0312-c3a0-4569-bea1-f076aee6e61f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.465014 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.465047 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/764b0312-c3a0-4569-bea1-f076aee6e61f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.465060 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6ptg\" (UniqueName: \"kubernetes.io/projected/764b0312-c3a0-4569-bea1-f076aee6e61f-kube-api-access-w6ptg\") on node \"crc\" DevicePath \"\"" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.767096 4789 generic.go:334] "Generic (PLEG): container finished" podID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerID="47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b" exitCode=0 Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.767155 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-srnl8" event={"ID":"764b0312-c3a0-4569-bea1-f076aee6e61f","Type":"ContainerDied","Data":"47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b"} Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.767188 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-srnl8" event={"ID":"764b0312-c3a0-4569-bea1-f076aee6e61f","Type":"ContainerDied","Data":"1c1860df98193471b88bdb10028724d4be39e0d54c95bea1f46a93a3840b4fae"} Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.767211 4789 scope.go:117] "RemoveContainer" containerID="47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.767241 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-srnl8" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.791998 4789 scope.go:117] "RemoveContainer" containerID="00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.824415 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-srnl8"] Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.825515 4789 scope.go:117] "RemoveContainer" containerID="4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.831576 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-srnl8"] Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.871875 4789 scope.go:117] "RemoveContainer" containerID="47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b" Oct 08 15:28:18 crc kubenswrapper[4789]: E1008 15:28:18.873369 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b\": container with ID starting with 47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b not found: ID does not exist" containerID="47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.873422 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b"} err="failed to get container status \"47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b\": rpc error: code = NotFound desc = could not find container \"47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b\": container with ID starting with 47e565079fbd627a4fae8522ec36b043ab7dbcee409d53d98070028abaabb32b not found: ID does not exist" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.873454 4789 scope.go:117] "RemoveContainer" containerID="00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429" Oct 08 15:28:18 crc kubenswrapper[4789]: E1008 15:28:18.873767 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429\": container with ID starting with 00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429 not found: ID does not exist" containerID="00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.873794 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429"} err="failed to get container status \"00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429\": rpc error: code = NotFound desc = could not find container \"00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429\": container with ID starting with 00942fd79f7238fd228704d90d753fd1bd63f655a6c066f6a8978fd747689429 not found: ID does not exist" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.873813 4789 scope.go:117] "RemoveContainer" containerID="4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d" Oct 08 15:28:18 crc kubenswrapper[4789]: E1008 15:28:18.874171 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d\": container with ID starting with 4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d not found: ID does not exist" containerID="4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d" Oct 08 15:28:18 crc kubenswrapper[4789]: I1008 15:28:18.874197 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d"} err="failed to get container status \"4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d\": rpc error: code = NotFound desc = could not find container \"4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d\": container with ID starting with 4273f122aa9738cbee4f8dbbe382e151a43fbe36265acb17664d034482f1928d not found: ID does not exist" Oct 08 15:28:20 crc kubenswrapper[4789]: I1008 15:28:20.741436 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" path="/var/lib/kubelet/pods/764b0312-c3a0-4569-bea1-f076aee6e61f/volumes" Oct 08 15:28:28 crc kubenswrapper[4789]: I1008 15:28:28.729647 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:28:28 crc kubenswrapper[4789]: E1008 15:28:28.730631 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:28:42 crc kubenswrapper[4789]: I1008 15:28:42.730403 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:28:42 crc kubenswrapper[4789]: E1008 15:28:42.731223 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:28:53 crc kubenswrapper[4789]: I1008 15:28:53.730217 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:28:53 crc kubenswrapper[4789]: E1008 15:28:53.730921 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:29:08 crc kubenswrapper[4789]: I1008 15:29:08.731366 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:29:08 crc kubenswrapper[4789]: E1008 15:29:08.732310 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:29:19 crc kubenswrapper[4789]: I1008 15:29:19.729790 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:29:19 crc kubenswrapper[4789]: E1008 15:29:19.730539 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:29:34 crc kubenswrapper[4789]: I1008 15:29:34.730837 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:29:34 crc kubenswrapper[4789]: E1008 15:29:34.732390 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:29:47 crc kubenswrapper[4789]: I1008 15:29:47.730277 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:29:47 crc kubenswrapper[4789]: E1008 15:29:47.731080 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:29:59 crc kubenswrapper[4789]: I1008 15:29:59.729773 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:29:59 crc kubenswrapper[4789]: E1008 15:29:59.731359 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.160018 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj"] Oct 08 15:30:00 crc kubenswrapper[4789]: E1008 15:30:00.160651 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="extract-utilities" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.160706 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="extract-utilities" Oct 08 15:30:00 crc kubenswrapper[4789]: E1008 15:30:00.160743 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="extract-content" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.160756 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="extract-content" Oct 08 15:30:00 crc kubenswrapper[4789]: E1008 15:30:00.160780 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="registry-server" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.160792 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="registry-server" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.161153 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="764b0312-c3a0-4569-bea1-f076aee6e61f" containerName="registry-server" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.162371 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.165774 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.165950 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.200115 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj"] Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.337281 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctvds\" (UniqueName: \"kubernetes.io/projected/e7dce57c-f414-4a02-8dec-18e1d9a8236b-kube-api-access-ctvds\") pod \"collect-profiles-29332290-6wklj\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.338228 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7dce57c-f414-4a02-8dec-18e1d9a8236b-config-volume\") pod \"collect-profiles-29332290-6wklj\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.338291 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7dce57c-f414-4a02-8dec-18e1d9a8236b-secret-volume\") pod \"collect-profiles-29332290-6wklj\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.440211 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctvds\" (UniqueName: \"kubernetes.io/projected/e7dce57c-f414-4a02-8dec-18e1d9a8236b-kube-api-access-ctvds\") pod \"collect-profiles-29332290-6wklj\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.440253 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7dce57c-f414-4a02-8dec-18e1d9a8236b-config-volume\") pod \"collect-profiles-29332290-6wklj\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.440277 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7dce57c-f414-4a02-8dec-18e1d9a8236b-secret-volume\") pod \"collect-profiles-29332290-6wklj\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.441415 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7dce57c-f414-4a02-8dec-18e1d9a8236b-config-volume\") pod \"collect-profiles-29332290-6wklj\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.446646 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7dce57c-f414-4a02-8dec-18e1d9a8236b-secret-volume\") pod \"collect-profiles-29332290-6wklj\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.458479 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctvds\" (UniqueName: \"kubernetes.io/projected/e7dce57c-f414-4a02-8dec-18e1d9a8236b-kube-api-access-ctvds\") pod \"collect-profiles-29332290-6wklj\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.484891 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:00 crc kubenswrapper[4789]: I1008 15:30:00.923210 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj"] Oct 08 15:30:01 crc kubenswrapper[4789]: I1008 15:30:01.778528 4789 generic.go:334] "Generic (PLEG): container finished" podID="e7dce57c-f414-4a02-8dec-18e1d9a8236b" containerID="e2f216ef11ef4ae9cabee57e45ebe566f5271f0f0f12e1bd90aae9e529625bfa" exitCode=0 Oct 08 15:30:01 crc kubenswrapper[4789]: I1008 15:30:01.778599 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" event={"ID":"e7dce57c-f414-4a02-8dec-18e1d9a8236b","Type":"ContainerDied","Data":"e2f216ef11ef4ae9cabee57e45ebe566f5271f0f0f12e1bd90aae9e529625bfa"} Oct 08 15:30:01 crc kubenswrapper[4789]: I1008 15:30:01.779083 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" event={"ID":"e7dce57c-f414-4a02-8dec-18e1d9a8236b","Type":"ContainerStarted","Data":"3aa132efcdee2daa07cda3f5b8bad3fd464be33f9f28d4749f332c078070f3ee"} Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.194884 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.293204 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctvds\" (UniqueName: \"kubernetes.io/projected/e7dce57c-f414-4a02-8dec-18e1d9a8236b-kube-api-access-ctvds\") pod \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.293529 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7dce57c-f414-4a02-8dec-18e1d9a8236b-config-volume\") pod \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.293666 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7dce57c-f414-4a02-8dec-18e1d9a8236b-secret-volume\") pod \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\" (UID: \"e7dce57c-f414-4a02-8dec-18e1d9a8236b\") " Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.294604 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7dce57c-f414-4a02-8dec-18e1d9a8236b-config-volume" (OuterVolumeSpecName: "config-volume") pod "e7dce57c-f414-4a02-8dec-18e1d9a8236b" (UID: "e7dce57c-f414-4a02-8dec-18e1d9a8236b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.306249 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7dce57c-f414-4a02-8dec-18e1d9a8236b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e7dce57c-f414-4a02-8dec-18e1d9a8236b" (UID: "e7dce57c-f414-4a02-8dec-18e1d9a8236b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.306352 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7dce57c-f414-4a02-8dec-18e1d9a8236b-kube-api-access-ctvds" (OuterVolumeSpecName: "kube-api-access-ctvds") pod "e7dce57c-f414-4a02-8dec-18e1d9a8236b" (UID: "e7dce57c-f414-4a02-8dec-18e1d9a8236b"). InnerVolumeSpecName "kube-api-access-ctvds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.395656 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7dce57c-f414-4a02-8dec-18e1d9a8236b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.395687 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctvds\" (UniqueName: \"kubernetes.io/projected/e7dce57c-f414-4a02-8dec-18e1d9a8236b-kube-api-access-ctvds\") on node \"crc\" DevicePath \"\"" Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.395696 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7dce57c-f414-4a02-8dec-18e1d9a8236b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.801297 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" event={"ID":"e7dce57c-f414-4a02-8dec-18e1d9a8236b","Type":"ContainerDied","Data":"3aa132efcdee2daa07cda3f5b8bad3fd464be33f9f28d4749f332c078070f3ee"} Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.801359 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332290-6wklj" Oct 08 15:30:03 crc kubenswrapper[4789]: I1008 15:30:03.801376 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3aa132efcdee2daa07cda3f5b8bad3fd464be33f9f28d4749f332c078070f3ee" Oct 08 15:30:04 crc kubenswrapper[4789]: I1008 15:30:04.275648 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb"] Oct 08 15:30:04 crc kubenswrapper[4789]: I1008 15:30:04.282908 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332245-fk2nb"] Oct 08 15:30:04 crc kubenswrapper[4789]: I1008 15:30:04.741557 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8edae8fb-3e38-477a-aa7f-5f6746833c47" path="/var/lib/kubelet/pods/8edae8fb-3e38-477a-aa7f-5f6746833c47/volumes" Oct 08 15:30:12 crc kubenswrapper[4789]: I1008 15:30:12.730182 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:30:12 crc kubenswrapper[4789]: E1008 15:30:12.730949 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:30:26 crc kubenswrapper[4789]: I1008 15:30:26.220484 4789 scope.go:117] "RemoveContainer" containerID="fb9bf00f8831ca940e98c3aadfe9d5641c3f9184fe4e80c4d7254b1e08190e2a" Oct 08 15:30:27 crc kubenswrapper[4789]: I1008 15:30:27.730605 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:30:28 crc kubenswrapper[4789]: I1008 15:30:28.008963 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"2548da8bb2ad7f23b4a07c4754dd1bd57a01e50746f4c72affc709ed1bec8c4e"} Oct 08 15:30:33 crc kubenswrapper[4789]: I1008 15:30:33.886773 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xxbg8"] Oct 08 15:30:33 crc kubenswrapper[4789]: E1008 15:30:33.887922 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7dce57c-f414-4a02-8dec-18e1d9a8236b" containerName="collect-profiles" Oct 08 15:30:33 crc kubenswrapper[4789]: I1008 15:30:33.887935 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7dce57c-f414-4a02-8dec-18e1d9a8236b" containerName="collect-profiles" Oct 08 15:30:33 crc kubenswrapper[4789]: I1008 15:30:33.888149 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7dce57c-f414-4a02-8dec-18e1d9a8236b" containerName="collect-profiles" Oct 08 15:30:33 crc kubenswrapper[4789]: I1008 15:30:33.889595 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:33 crc kubenswrapper[4789]: I1008 15:30:33.910512 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxbg8"] Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.024675 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fprvg\" (UniqueName: \"kubernetes.io/projected/e96c5bc2-36d7-4194-bec1-f176c76a6de7-kube-api-access-fprvg\") pod \"redhat-marketplace-xxbg8\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.024735 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-utilities\") pod \"redhat-marketplace-xxbg8\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.024850 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-catalog-content\") pod \"redhat-marketplace-xxbg8\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.126919 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-catalog-content\") pod \"redhat-marketplace-xxbg8\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.127151 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fprvg\" (UniqueName: \"kubernetes.io/projected/e96c5bc2-36d7-4194-bec1-f176c76a6de7-kube-api-access-fprvg\") pod \"redhat-marketplace-xxbg8\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.127198 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-utilities\") pod \"redhat-marketplace-xxbg8\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.127877 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-catalog-content\") pod \"redhat-marketplace-xxbg8\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.127937 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-utilities\") pod \"redhat-marketplace-xxbg8\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.158720 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fprvg\" (UniqueName: \"kubernetes.io/projected/e96c5bc2-36d7-4194-bec1-f176c76a6de7-kube-api-access-fprvg\") pod \"redhat-marketplace-xxbg8\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.228474 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:34 crc kubenswrapper[4789]: I1008 15:30:34.707282 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxbg8"] Oct 08 15:30:35 crc kubenswrapper[4789]: I1008 15:30:35.075205 4789 generic.go:334] "Generic (PLEG): container finished" podID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerID="0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785" exitCode=0 Oct 08 15:30:35 crc kubenswrapper[4789]: I1008 15:30:35.075314 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxbg8" event={"ID":"e96c5bc2-36d7-4194-bec1-f176c76a6de7","Type":"ContainerDied","Data":"0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785"} Oct 08 15:30:35 crc kubenswrapper[4789]: I1008 15:30:35.075471 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxbg8" event={"ID":"e96c5bc2-36d7-4194-bec1-f176c76a6de7","Type":"ContainerStarted","Data":"be5f5ab0e7052f40f2ab6f1640e46cf9081995be5e1329d99b14fec5e8e9a9dd"} Oct 08 15:30:35 crc kubenswrapper[4789]: I1008 15:30:35.076880 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:30:37 crc kubenswrapper[4789]: I1008 15:30:37.102481 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxbg8" event={"ID":"e96c5bc2-36d7-4194-bec1-f176c76a6de7","Type":"ContainerStarted","Data":"b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368"} Oct 08 15:30:38 crc kubenswrapper[4789]: I1008 15:30:38.113200 4789 generic.go:334] "Generic (PLEG): container finished" podID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerID="b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368" exitCode=0 Oct 08 15:30:38 crc kubenswrapper[4789]: I1008 15:30:38.113299 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxbg8" event={"ID":"e96c5bc2-36d7-4194-bec1-f176c76a6de7","Type":"ContainerDied","Data":"b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368"} Oct 08 15:30:39 crc kubenswrapper[4789]: I1008 15:30:39.126855 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxbg8" event={"ID":"e96c5bc2-36d7-4194-bec1-f176c76a6de7","Type":"ContainerStarted","Data":"ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861"} Oct 08 15:30:39 crc kubenswrapper[4789]: I1008 15:30:39.155685 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xxbg8" podStartSLOduration=2.73801517 podStartE2EDuration="6.155665559s" podCreationTimestamp="2025-10-08 15:30:33 +0000 UTC" firstStartedPulling="2025-10-08 15:30:35.076604421 +0000 UTC m=+5374.983351913" lastFinishedPulling="2025-10-08 15:30:38.49425481 +0000 UTC m=+5378.401002302" observedRunningTime="2025-10-08 15:30:39.153324795 +0000 UTC m=+5379.060072287" watchObservedRunningTime="2025-10-08 15:30:39.155665559 +0000 UTC m=+5379.062413051" Oct 08 15:30:44 crc kubenswrapper[4789]: I1008 15:30:44.228926 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:44 crc kubenswrapper[4789]: I1008 15:30:44.229436 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:44 crc kubenswrapper[4789]: I1008 15:30:44.272126 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:45 crc kubenswrapper[4789]: I1008 15:30:45.250922 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:45 crc kubenswrapper[4789]: I1008 15:30:45.306138 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxbg8"] Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.213081 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xxbg8" podUID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerName="registry-server" containerID="cri-o://ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861" gracePeriod=2 Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.734489 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.892704 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fprvg\" (UniqueName: \"kubernetes.io/projected/e96c5bc2-36d7-4194-bec1-f176c76a6de7-kube-api-access-fprvg\") pod \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.892751 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-utilities\") pod \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.892773 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-catalog-content\") pod \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\" (UID: \"e96c5bc2-36d7-4194-bec1-f176c76a6de7\") " Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.894340 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-utilities" (OuterVolumeSpecName: "utilities") pod "e96c5bc2-36d7-4194-bec1-f176c76a6de7" (UID: "e96c5bc2-36d7-4194-bec1-f176c76a6de7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.898748 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e96c5bc2-36d7-4194-bec1-f176c76a6de7-kube-api-access-fprvg" (OuterVolumeSpecName: "kube-api-access-fprvg") pod "e96c5bc2-36d7-4194-bec1-f176c76a6de7" (UID: "e96c5bc2-36d7-4194-bec1-f176c76a6de7"). InnerVolumeSpecName "kube-api-access-fprvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.909326 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e96c5bc2-36d7-4194-bec1-f176c76a6de7" (UID: "e96c5bc2-36d7-4194-bec1-f176c76a6de7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.995106 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fprvg\" (UniqueName: \"kubernetes.io/projected/e96c5bc2-36d7-4194-bec1-f176c76a6de7-kube-api-access-fprvg\") on node \"crc\" DevicePath \"\"" Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.995147 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:30:47 crc kubenswrapper[4789]: I1008 15:30:47.995159 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e96c5bc2-36d7-4194-bec1-f176c76a6de7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.225866 4789 generic.go:334] "Generic (PLEG): container finished" podID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerID="ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861" exitCode=0 Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.225910 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxbg8" event={"ID":"e96c5bc2-36d7-4194-bec1-f176c76a6de7","Type":"ContainerDied","Data":"ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861"} Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.225938 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxbg8" event={"ID":"e96c5bc2-36d7-4194-bec1-f176c76a6de7","Type":"ContainerDied","Data":"be5f5ab0e7052f40f2ab6f1640e46cf9081995be5e1329d99b14fec5e8e9a9dd"} Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.225957 4789 scope.go:117] "RemoveContainer" containerID="ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.225961 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxbg8" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.259898 4789 scope.go:117] "RemoveContainer" containerID="b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.263241 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxbg8"] Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.283372 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxbg8"] Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.283674 4789 scope.go:117] "RemoveContainer" containerID="0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.334857 4789 scope.go:117] "RemoveContainer" containerID="ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861" Oct 08 15:30:48 crc kubenswrapper[4789]: E1008 15:30:48.335576 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861\": container with ID starting with ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861 not found: ID does not exist" containerID="ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.335621 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861"} err="failed to get container status \"ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861\": rpc error: code = NotFound desc = could not find container \"ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861\": container with ID starting with ed1f06787fe5f898219ad862d7f20a118f395522a6e86eec9f3de1cbc3fc1861 not found: ID does not exist" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.335643 4789 scope.go:117] "RemoveContainer" containerID="b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368" Oct 08 15:30:48 crc kubenswrapper[4789]: E1008 15:30:48.336079 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368\": container with ID starting with b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368 not found: ID does not exist" containerID="b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.336098 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368"} err="failed to get container status \"b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368\": rpc error: code = NotFound desc = could not find container \"b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368\": container with ID starting with b6d5dce8e9bed073146225d774d14c6d8ecd050b7848f0a1add7744d2ae7c368 not found: ID does not exist" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.336111 4789 scope.go:117] "RemoveContainer" containerID="0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785" Oct 08 15:30:48 crc kubenswrapper[4789]: E1008 15:30:48.336429 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785\": container with ID starting with 0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785 not found: ID does not exist" containerID="0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.336448 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785"} err="failed to get container status \"0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785\": rpc error: code = NotFound desc = could not find container \"0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785\": container with ID starting with 0d631d11aa0c3a63717053932b181ff90c4b03c4f53f5fa575b7c5f1b53f3785 not found: ID does not exist" Oct 08 15:30:48 crc kubenswrapper[4789]: I1008 15:30:48.741703 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" path="/var/lib/kubelet/pods/e96c5bc2-36d7-4194-bec1-f176c76a6de7/volumes" Oct 08 15:32:56 crc kubenswrapper[4789]: I1008 15:32:56.433108 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:32:56 crc kubenswrapper[4789]: I1008 15:32:56.433624 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:33:26 crc kubenswrapper[4789]: I1008 15:33:26.432548 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:33:26 crc kubenswrapper[4789]: I1008 15:33:26.433143 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:33:56 crc kubenswrapper[4789]: I1008 15:33:56.432455 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:33:56 crc kubenswrapper[4789]: I1008 15:33:56.433078 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:33:56 crc kubenswrapper[4789]: I1008 15:33:56.433121 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 15:33:56 crc kubenswrapper[4789]: I1008 15:33:56.433769 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2548da8bb2ad7f23b4a07c4754dd1bd57a01e50746f4c72affc709ed1bec8c4e"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:33:56 crc kubenswrapper[4789]: I1008 15:33:56.433830 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://2548da8bb2ad7f23b4a07c4754dd1bd57a01e50746f4c72affc709ed1bec8c4e" gracePeriod=600 Oct 08 15:33:56 crc kubenswrapper[4789]: I1008 15:33:56.967933 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="2548da8bb2ad7f23b4a07c4754dd1bd57a01e50746f4c72affc709ed1bec8c4e" exitCode=0 Oct 08 15:33:56 crc kubenswrapper[4789]: I1008 15:33:56.968026 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"2548da8bb2ad7f23b4a07c4754dd1bd57a01e50746f4c72affc709ed1bec8c4e"} Oct 08 15:33:56 crc kubenswrapper[4789]: I1008 15:33:56.968370 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a"} Oct 08 15:33:56 crc kubenswrapper[4789]: I1008 15:33:56.968398 4789 scope.go:117] "RemoveContainer" containerID="8152515af403a8537c2b4d3bb43839754d12f06872522c25045a67bc38f156c4" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.636036 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dwhls"] Oct 08 15:34:19 crc kubenswrapper[4789]: E1008 15:34:19.637181 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerName="extract-utilities" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.637198 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerName="extract-utilities" Oct 08 15:34:19 crc kubenswrapper[4789]: E1008 15:34:19.637217 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerName="extract-content" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.637225 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerName="extract-content" Oct 08 15:34:19 crc kubenswrapper[4789]: E1008 15:34:19.637247 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerName="registry-server" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.637255 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerName="registry-server" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.637519 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e96c5bc2-36d7-4194-bec1-f176c76a6de7" containerName="registry-server" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.639558 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.659793 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dwhls"] Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.749491 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxxlh\" (UniqueName: \"kubernetes.io/projected/a3585cf0-1088-4263-ae40-b7a3998399a8-kube-api-access-qxxlh\") pod \"community-operators-dwhls\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.749647 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-catalog-content\") pod \"community-operators-dwhls\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.749780 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-utilities\") pod \"community-operators-dwhls\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.851632 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxxlh\" (UniqueName: \"kubernetes.io/projected/a3585cf0-1088-4263-ae40-b7a3998399a8-kube-api-access-qxxlh\") pod \"community-operators-dwhls\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.851822 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-catalog-content\") pod \"community-operators-dwhls\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.851891 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-utilities\") pod \"community-operators-dwhls\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.852382 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-catalog-content\") pod \"community-operators-dwhls\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.852415 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-utilities\") pod \"community-operators-dwhls\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.872026 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxxlh\" (UniqueName: \"kubernetes.io/projected/a3585cf0-1088-4263-ae40-b7a3998399a8-kube-api-access-qxxlh\") pod \"community-operators-dwhls\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:19 crc kubenswrapper[4789]: I1008 15:34:19.972233 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:20 crc kubenswrapper[4789]: I1008 15:34:20.564047 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dwhls"] Oct 08 15:34:21 crc kubenswrapper[4789]: I1008 15:34:21.226334 4789 generic.go:334] "Generic (PLEG): container finished" podID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerID="16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01" exitCode=0 Oct 08 15:34:21 crc kubenswrapper[4789]: I1008 15:34:21.226420 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwhls" event={"ID":"a3585cf0-1088-4263-ae40-b7a3998399a8","Type":"ContainerDied","Data":"16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01"} Oct 08 15:34:21 crc kubenswrapper[4789]: I1008 15:34:21.226555 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwhls" event={"ID":"a3585cf0-1088-4263-ae40-b7a3998399a8","Type":"ContainerStarted","Data":"2c3323f11fb8a6cf505d1c8e4abfcda7266927bf5612cbbcca469e670e2cb0cd"} Oct 08 15:34:23 crc kubenswrapper[4789]: I1008 15:34:23.250302 4789 generic.go:334] "Generic (PLEG): container finished" podID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerID="15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da" exitCode=0 Oct 08 15:34:23 crc kubenswrapper[4789]: I1008 15:34:23.250798 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwhls" event={"ID":"a3585cf0-1088-4263-ae40-b7a3998399a8","Type":"ContainerDied","Data":"15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da"} Oct 08 15:34:24 crc kubenswrapper[4789]: I1008 15:34:24.261016 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwhls" event={"ID":"a3585cf0-1088-4263-ae40-b7a3998399a8","Type":"ContainerStarted","Data":"6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c"} Oct 08 15:34:24 crc kubenswrapper[4789]: I1008 15:34:24.286283 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dwhls" podStartSLOduration=2.768039461 podStartE2EDuration="5.286268364s" podCreationTimestamp="2025-10-08 15:34:19 +0000 UTC" firstStartedPulling="2025-10-08 15:34:21.228962727 +0000 UTC m=+5601.135710239" lastFinishedPulling="2025-10-08 15:34:23.74719165 +0000 UTC m=+5603.653939142" observedRunningTime="2025-10-08 15:34:24.279971573 +0000 UTC m=+5604.186719065" watchObservedRunningTime="2025-10-08 15:34:24.286268364 +0000 UTC m=+5604.193015856" Oct 08 15:34:29 crc kubenswrapper[4789]: I1008 15:34:29.972442 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:29 crc kubenswrapper[4789]: I1008 15:34:29.972998 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:30 crc kubenswrapper[4789]: I1008 15:34:30.040404 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:30 crc kubenswrapper[4789]: I1008 15:34:30.357475 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:30 crc kubenswrapper[4789]: I1008 15:34:30.402826 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dwhls"] Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.327214 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dwhls" podUID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerName="registry-server" containerID="cri-o://6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c" gracePeriod=2 Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.852342 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.892147 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-catalog-content\") pod \"a3585cf0-1088-4263-ae40-b7a3998399a8\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.892408 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxxlh\" (UniqueName: \"kubernetes.io/projected/a3585cf0-1088-4263-ae40-b7a3998399a8-kube-api-access-qxxlh\") pod \"a3585cf0-1088-4263-ae40-b7a3998399a8\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.892462 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-utilities\") pod \"a3585cf0-1088-4263-ae40-b7a3998399a8\" (UID: \"a3585cf0-1088-4263-ae40-b7a3998399a8\") " Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.894287 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-utilities" (OuterVolumeSpecName: "utilities") pod "a3585cf0-1088-4263-ae40-b7a3998399a8" (UID: "a3585cf0-1088-4263-ae40-b7a3998399a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.912380 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3585cf0-1088-4263-ae40-b7a3998399a8-kube-api-access-qxxlh" (OuterVolumeSpecName: "kube-api-access-qxxlh") pod "a3585cf0-1088-4263-ae40-b7a3998399a8" (UID: "a3585cf0-1088-4263-ae40-b7a3998399a8"). InnerVolumeSpecName "kube-api-access-qxxlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.945155 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3585cf0-1088-4263-ae40-b7a3998399a8" (UID: "a3585cf0-1088-4263-ae40-b7a3998399a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.994108 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.994425 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxxlh\" (UniqueName: \"kubernetes.io/projected/a3585cf0-1088-4263-ae40-b7a3998399a8-kube-api-access-qxxlh\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:32 crc kubenswrapper[4789]: I1008 15:34:32.994500 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3585cf0-1088-4263-ae40-b7a3998399a8-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.336546 4789 generic.go:334] "Generic (PLEG): container finished" podID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerID="6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c" exitCode=0 Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.336606 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwhls" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.336627 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwhls" event={"ID":"a3585cf0-1088-4263-ae40-b7a3998399a8","Type":"ContainerDied","Data":"6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c"} Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.340154 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwhls" event={"ID":"a3585cf0-1088-4263-ae40-b7a3998399a8","Type":"ContainerDied","Data":"2c3323f11fb8a6cf505d1c8e4abfcda7266927bf5612cbbcca469e670e2cb0cd"} Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.340188 4789 scope.go:117] "RemoveContainer" containerID="6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.368048 4789 scope.go:117] "RemoveContainer" containerID="15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.380678 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dwhls"] Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.389678 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dwhls"] Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.395663 4789 scope.go:117] "RemoveContainer" containerID="16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.442140 4789 scope.go:117] "RemoveContainer" containerID="6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c" Oct 08 15:34:33 crc kubenswrapper[4789]: E1008 15:34:33.442524 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c\": container with ID starting with 6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c not found: ID does not exist" containerID="6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.442556 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c"} err="failed to get container status \"6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c\": rpc error: code = NotFound desc = could not find container \"6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c\": container with ID starting with 6db2e38108cbd279bd20d6c86f58b98b3d78359e1df4f2c0eb8f4aa014775e9c not found: ID does not exist" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.442580 4789 scope.go:117] "RemoveContainer" containerID="15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da" Oct 08 15:34:33 crc kubenswrapper[4789]: E1008 15:34:33.442882 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da\": container with ID starting with 15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da not found: ID does not exist" containerID="15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.442908 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da"} err="failed to get container status \"15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da\": rpc error: code = NotFound desc = could not find container \"15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da\": container with ID starting with 15e6f4bfbd7e20f9e786517b0520e91d7eb3bb259c6913aec0569059488510da not found: ID does not exist" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.442924 4789 scope.go:117] "RemoveContainer" containerID="16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01" Oct 08 15:34:33 crc kubenswrapper[4789]: E1008 15:34:33.443218 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01\": container with ID starting with 16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01 not found: ID does not exist" containerID="16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01" Oct 08 15:34:33 crc kubenswrapper[4789]: I1008 15:34:33.443242 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01"} err="failed to get container status \"16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01\": rpc error: code = NotFound desc = could not find container \"16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01\": container with ID starting with 16fa8084d2404eab4507a7a71054a5e7ec7fba378ab92f23aa2f3671bd1f3e01 not found: ID does not exist" Oct 08 15:34:34 crc kubenswrapper[4789]: I1008 15:34:34.741027 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3585cf0-1088-4263-ae40-b7a3998399a8" path="/var/lib/kubelet/pods/a3585cf0-1088-4263-ae40-b7a3998399a8/volumes" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.782657 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qw86x"] Oct 08 15:34:41 crc kubenswrapper[4789]: E1008 15:34:41.783581 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerName="extract-content" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.783601 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerName="extract-content" Oct 08 15:34:41 crc kubenswrapper[4789]: E1008 15:34:41.783628 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerName="extract-utilities" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.783637 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerName="extract-utilities" Oct 08 15:34:41 crc kubenswrapper[4789]: E1008 15:34:41.783670 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerName="registry-server" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.783678 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerName="registry-server" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.783942 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3585cf0-1088-4263-ae40-b7a3998399a8" containerName="registry-server" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.786280 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.792426 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qw86x"] Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.877528 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-utilities\") pod \"certified-operators-qw86x\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.877847 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wbgm\" (UniqueName: \"kubernetes.io/projected/0911be01-722d-4292-a963-041fe2ef59dc-kube-api-access-2wbgm\") pod \"certified-operators-qw86x\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.877928 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-catalog-content\") pod \"certified-operators-qw86x\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.980580 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wbgm\" (UniqueName: \"kubernetes.io/projected/0911be01-722d-4292-a963-041fe2ef59dc-kube-api-access-2wbgm\") pod \"certified-operators-qw86x\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.980686 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-catalog-content\") pod \"certified-operators-qw86x\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.980777 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-utilities\") pod \"certified-operators-qw86x\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.981279 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-utilities\") pod \"certified-operators-qw86x\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:41 crc kubenswrapper[4789]: I1008 15:34:41.981335 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-catalog-content\") pod \"certified-operators-qw86x\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:42 crc kubenswrapper[4789]: I1008 15:34:42.007968 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wbgm\" (UniqueName: \"kubernetes.io/projected/0911be01-722d-4292-a963-041fe2ef59dc-kube-api-access-2wbgm\") pod \"certified-operators-qw86x\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:42 crc kubenswrapper[4789]: I1008 15:34:42.118722 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:42 crc kubenswrapper[4789]: I1008 15:34:42.516742 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qw86x"] Oct 08 15:34:43 crc kubenswrapper[4789]: I1008 15:34:43.439512 4789 generic.go:334] "Generic (PLEG): container finished" podID="0911be01-722d-4292-a963-041fe2ef59dc" containerID="a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea" exitCode=0 Oct 08 15:34:43 crc kubenswrapper[4789]: I1008 15:34:43.439571 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw86x" event={"ID":"0911be01-722d-4292-a963-041fe2ef59dc","Type":"ContainerDied","Data":"a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea"} Oct 08 15:34:43 crc kubenswrapper[4789]: I1008 15:34:43.440123 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw86x" event={"ID":"0911be01-722d-4292-a963-041fe2ef59dc","Type":"ContainerStarted","Data":"13ccf7dd345239065e9ec574fed8ff90f79d644769ab830fd45210885f9b9a78"} Oct 08 15:34:44 crc kubenswrapper[4789]: I1008 15:34:44.449971 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw86x" event={"ID":"0911be01-722d-4292-a963-041fe2ef59dc","Type":"ContainerStarted","Data":"af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d"} Oct 08 15:34:45 crc kubenswrapper[4789]: I1008 15:34:45.459894 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw86x" event={"ID":"0911be01-722d-4292-a963-041fe2ef59dc","Type":"ContainerDied","Data":"af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d"} Oct 08 15:34:45 crc kubenswrapper[4789]: I1008 15:34:45.459890 4789 generic.go:334] "Generic (PLEG): container finished" podID="0911be01-722d-4292-a963-041fe2ef59dc" containerID="af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d" exitCode=0 Oct 08 15:34:46 crc kubenswrapper[4789]: I1008 15:34:46.473501 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw86x" event={"ID":"0911be01-722d-4292-a963-041fe2ef59dc","Type":"ContainerStarted","Data":"1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5"} Oct 08 15:34:52 crc kubenswrapper[4789]: I1008 15:34:52.119652 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:52 crc kubenswrapper[4789]: I1008 15:34:52.120273 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:52 crc kubenswrapper[4789]: I1008 15:34:52.179395 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:52 crc kubenswrapper[4789]: I1008 15:34:52.206503 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qw86x" podStartSLOduration=8.494383065 podStartE2EDuration="11.206483079s" podCreationTimestamp="2025-10-08 15:34:41 +0000 UTC" firstStartedPulling="2025-10-08 15:34:43.44243039 +0000 UTC m=+5623.349177882" lastFinishedPulling="2025-10-08 15:34:46.154530404 +0000 UTC m=+5626.061277896" observedRunningTime="2025-10-08 15:34:46.500303722 +0000 UTC m=+5626.407051214" watchObservedRunningTime="2025-10-08 15:34:52.206483079 +0000 UTC m=+5632.113230571" Oct 08 15:34:52 crc kubenswrapper[4789]: I1008 15:34:52.571648 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:52 crc kubenswrapper[4789]: I1008 15:34:52.622232 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qw86x"] Oct 08 15:34:54 crc kubenswrapper[4789]: I1008 15:34:54.544272 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qw86x" podUID="0911be01-722d-4292-a963-041fe2ef59dc" containerName="registry-server" containerID="cri-o://1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5" gracePeriod=2 Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.064959 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.153429 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wbgm\" (UniqueName: \"kubernetes.io/projected/0911be01-722d-4292-a963-041fe2ef59dc-kube-api-access-2wbgm\") pod \"0911be01-722d-4292-a963-041fe2ef59dc\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.153796 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-utilities\") pod \"0911be01-722d-4292-a963-041fe2ef59dc\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.154170 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-catalog-content\") pod \"0911be01-722d-4292-a963-041fe2ef59dc\" (UID: \"0911be01-722d-4292-a963-041fe2ef59dc\") " Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.154614 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-utilities" (OuterVolumeSpecName: "utilities") pod "0911be01-722d-4292-a963-041fe2ef59dc" (UID: "0911be01-722d-4292-a963-041fe2ef59dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.155039 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.159835 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0911be01-722d-4292-a963-041fe2ef59dc-kube-api-access-2wbgm" (OuterVolumeSpecName: "kube-api-access-2wbgm") pod "0911be01-722d-4292-a963-041fe2ef59dc" (UID: "0911be01-722d-4292-a963-041fe2ef59dc"). InnerVolumeSpecName "kube-api-access-2wbgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.217958 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0911be01-722d-4292-a963-041fe2ef59dc" (UID: "0911be01-722d-4292-a963-041fe2ef59dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.257323 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0911be01-722d-4292-a963-041fe2ef59dc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.257359 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wbgm\" (UniqueName: \"kubernetes.io/projected/0911be01-722d-4292-a963-041fe2ef59dc-kube-api-access-2wbgm\") on node \"crc\" DevicePath \"\"" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.558267 4789 generic.go:334] "Generic (PLEG): container finished" podID="0911be01-722d-4292-a963-041fe2ef59dc" containerID="1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5" exitCode=0 Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.558338 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw86x" event={"ID":"0911be01-722d-4292-a963-041fe2ef59dc","Type":"ContainerDied","Data":"1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5"} Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.558662 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qw86x" event={"ID":"0911be01-722d-4292-a963-041fe2ef59dc","Type":"ContainerDied","Data":"13ccf7dd345239065e9ec574fed8ff90f79d644769ab830fd45210885f9b9a78"} Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.558683 4789 scope.go:117] "RemoveContainer" containerID="1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.558374 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qw86x" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.585366 4789 scope.go:117] "RemoveContainer" containerID="af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.612717 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qw86x"] Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.620426 4789 scope.go:117] "RemoveContainer" containerID="a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.620923 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qw86x"] Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.667817 4789 scope.go:117] "RemoveContainer" containerID="1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5" Oct 08 15:34:55 crc kubenswrapper[4789]: E1008 15:34:55.668381 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5\": container with ID starting with 1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5 not found: ID does not exist" containerID="1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.668480 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5"} err="failed to get container status \"1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5\": rpc error: code = NotFound desc = could not find container \"1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5\": container with ID starting with 1076e4a131cc836a2147d5ff6fb0c98425c2c24bafc91d8e969d64ed1bf61ea5 not found: ID does not exist" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.668547 4789 scope.go:117] "RemoveContainer" containerID="af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d" Oct 08 15:34:55 crc kubenswrapper[4789]: E1008 15:34:55.668998 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d\": container with ID starting with af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d not found: ID does not exist" containerID="af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.669039 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d"} err="failed to get container status \"af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d\": rpc error: code = NotFound desc = could not find container \"af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d\": container with ID starting with af8cc931048b4057e1ed2c906c6a757d71b276064ae985d518034dcd673a8f3d not found: ID does not exist" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.669070 4789 scope.go:117] "RemoveContainer" containerID="a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea" Oct 08 15:34:55 crc kubenswrapper[4789]: E1008 15:34:55.669429 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea\": container with ID starting with a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea not found: ID does not exist" containerID="a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea" Oct 08 15:34:55 crc kubenswrapper[4789]: I1008 15:34:55.669489 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea"} err="failed to get container status \"a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea\": rpc error: code = NotFound desc = could not find container \"a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea\": container with ID starting with a879c5f8feac265ac27d9c255cac307c2da0481ad248c48a6518d4796659f5ea not found: ID does not exist" Oct 08 15:34:56 crc kubenswrapper[4789]: I1008 15:34:56.742665 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0911be01-722d-4292-a963-041fe2ef59dc" path="/var/lib/kubelet/pods/0911be01-722d-4292-a963-041fe2ef59dc/volumes" Oct 08 15:35:56 crc kubenswrapper[4789]: I1008 15:35:56.433617 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:35:56 crc kubenswrapper[4789]: I1008 15:35:56.434666 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:36:26 crc kubenswrapper[4789]: I1008 15:36:26.432571 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:36:26 crc kubenswrapper[4789]: I1008 15:36:26.433144 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:36:56 crc kubenswrapper[4789]: I1008 15:36:56.433131 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:36:56 crc kubenswrapper[4789]: I1008 15:36:56.433742 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:36:56 crc kubenswrapper[4789]: I1008 15:36:56.433795 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 15:36:56 crc kubenswrapper[4789]: I1008 15:36:56.434600 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:36:56 crc kubenswrapper[4789]: I1008 15:36:56.434647 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" gracePeriod=600 Oct 08 15:36:56 crc kubenswrapper[4789]: E1008 15:36:56.556434 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:36:56 crc kubenswrapper[4789]: I1008 15:36:56.704411 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" exitCode=0 Oct 08 15:36:56 crc kubenswrapper[4789]: I1008 15:36:56.704475 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a"} Oct 08 15:36:56 crc kubenswrapper[4789]: I1008 15:36:56.704758 4789 scope.go:117] "RemoveContainer" containerID="2548da8bb2ad7f23b4a07c4754dd1bd57a01e50746f4c72affc709ed1bec8c4e" Oct 08 15:36:56 crc kubenswrapper[4789]: I1008 15:36:56.705686 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:36:56 crc kubenswrapper[4789]: E1008 15:36:56.706110 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:37:07 crc kubenswrapper[4789]: I1008 15:37:07.730513 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:37:07 crc kubenswrapper[4789]: E1008 15:37:07.731469 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:37:16 crc kubenswrapper[4789]: I1008 15:37:16.898072 4789 generic.go:334] "Generic (PLEG): container finished" podID="5622aa64-be18-4420-8c84-b76fe03c1346" containerID="84e3bd795069144ca408b979b068ed5ffe872cc3f6b473b080b59c236d61ecb4" exitCode=0 Oct 08 15:37:16 crc kubenswrapper[4789]: I1008 15:37:16.898166 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5622aa64-be18-4420-8c84-b76fe03c1346","Type":"ContainerDied","Data":"84e3bd795069144ca408b979b068ed5ffe872cc3f6b473b080b59c236d61ecb4"} Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.309676 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.421179 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config-secret\") pod \"5622aa64-be18-4420-8c84-b76fe03c1346\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.421562 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-workdir\") pod \"5622aa64-be18-4420-8c84-b76fe03c1346\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.421647 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ssh-key\") pod \"5622aa64-be18-4420-8c84-b76fe03c1346\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.421707 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-temporary\") pod \"5622aa64-be18-4420-8c84-b76fe03c1346\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.421734 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"5622aa64-be18-4420-8c84-b76fe03c1346\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.421833 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-config-data\") pod \"5622aa64-be18-4420-8c84-b76fe03c1346\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.421886 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ca-certs\") pod \"5622aa64-be18-4420-8c84-b76fe03c1346\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.421910 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpd2s\" (UniqueName: \"kubernetes.io/projected/5622aa64-be18-4420-8c84-b76fe03c1346-kube-api-access-zpd2s\") pod \"5622aa64-be18-4420-8c84-b76fe03c1346\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.421960 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config\") pod \"5622aa64-be18-4420-8c84-b76fe03c1346\" (UID: \"5622aa64-be18-4420-8c84-b76fe03c1346\") " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.422374 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "5622aa64-be18-4420-8c84-b76fe03c1346" (UID: "5622aa64-be18-4420-8c84-b76fe03c1346"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.422671 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.422675 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-config-data" (OuterVolumeSpecName: "config-data") pod "5622aa64-be18-4420-8c84-b76fe03c1346" (UID: "5622aa64-be18-4420-8c84-b76fe03c1346"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.431497 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "5622aa64-be18-4420-8c84-b76fe03c1346" (UID: "5622aa64-be18-4420-8c84-b76fe03c1346"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.432261 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5622aa64-be18-4420-8c84-b76fe03c1346-kube-api-access-zpd2s" (OuterVolumeSpecName: "kube-api-access-zpd2s") pod "5622aa64-be18-4420-8c84-b76fe03c1346" (UID: "5622aa64-be18-4420-8c84-b76fe03c1346"). InnerVolumeSpecName "kube-api-access-zpd2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.437122 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "5622aa64-be18-4420-8c84-b76fe03c1346" (UID: "5622aa64-be18-4420-8c84-b76fe03c1346"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.456751 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "5622aa64-be18-4420-8c84-b76fe03c1346" (UID: "5622aa64-be18-4420-8c84-b76fe03c1346"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.459554 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5622aa64-be18-4420-8c84-b76fe03c1346" (UID: "5622aa64-be18-4420-8c84-b76fe03c1346"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.472177 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5622aa64-be18-4420-8c84-b76fe03c1346" (UID: "5622aa64-be18-4420-8c84-b76fe03c1346"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.485598 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5622aa64-be18-4420-8c84-b76fe03c1346" (UID: "5622aa64-be18-4420-8c84-b76fe03c1346"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.524666 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpd2s\" (UniqueName: \"kubernetes.io/projected/5622aa64-be18-4420-8c84-b76fe03c1346-kube-api-access-zpd2s\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.524710 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.524727 4789 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.524742 4789 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5622aa64-be18-4420-8c84-b76fe03c1346-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.524756 4789 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.524801 4789 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.524814 4789 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5622aa64-be18-4420-8c84-b76fe03c1346-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.524826 4789 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5622aa64-be18-4420-8c84-b76fe03c1346-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.552920 4789 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.626342 4789 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.730386 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:37:18 crc kubenswrapper[4789]: E1008 15:37:18.730819 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.918884 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5622aa64-be18-4420-8c84-b76fe03c1346","Type":"ContainerDied","Data":"f18c299b113e7a48a2c3dfe5a9dc13a9ad3198cd51144a885c971a52a8934235"} Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.918926 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f18c299b113e7a48a2c3dfe5a9dc13a9ad3198cd51144a885c971a52a8934235" Oct 08 15:37:18 crc kubenswrapper[4789]: I1008 15:37:18.918947 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.559184 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 15:37:30 crc kubenswrapper[4789]: E1008 15:37:30.560126 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0911be01-722d-4292-a963-041fe2ef59dc" containerName="extract-utilities" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.560139 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0911be01-722d-4292-a963-041fe2ef59dc" containerName="extract-utilities" Oct 08 15:37:30 crc kubenswrapper[4789]: E1008 15:37:30.560156 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0911be01-722d-4292-a963-041fe2ef59dc" containerName="extract-content" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.560163 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0911be01-722d-4292-a963-041fe2ef59dc" containerName="extract-content" Oct 08 15:37:30 crc kubenswrapper[4789]: E1008 15:37:30.560201 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5622aa64-be18-4420-8c84-b76fe03c1346" containerName="tempest-tests-tempest-tests-runner" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.560208 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="5622aa64-be18-4420-8c84-b76fe03c1346" containerName="tempest-tests-tempest-tests-runner" Oct 08 15:37:30 crc kubenswrapper[4789]: E1008 15:37:30.560244 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0911be01-722d-4292-a963-041fe2ef59dc" containerName="registry-server" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.560250 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="0911be01-722d-4292-a963-041fe2ef59dc" containerName="registry-server" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.560444 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="5622aa64-be18-4420-8c84-b76fe03c1346" containerName="tempest-tests-tempest-tests-runner" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.560458 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="0911be01-722d-4292-a963-041fe2ef59dc" containerName="registry-server" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.561322 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.563417 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-thkkz" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.575092 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.626610 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5bgs\" (UniqueName: \"kubernetes.io/projected/3d01bf45-7cf8-44f1-a306-8ed97d937dfe-kube-api-access-w5bgs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d01bf45-7cf8-44f1-a306-8ed97d937dfe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.626716 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d01bf45-7cf8-44f1-a306-8ed97d937dfe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.732606 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5bgs\" (UniqueName: \"kubernetes.io/projected/3d01bf45-7cf8-44f1-a306-8ed97d937dfe-kube-api-access-w5bgs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d01bf45-7cf8-44f1-a306-8ed97d937dfe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.732976 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d01bf45-7cf8-44f1-a306-8ed97d937dfe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.739752 4789 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d01bf45-7cf8-44f1-a306-8ed97d937dfe\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.777657 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5bgs\" (UniqueName: \"kubernetes.io/projected/3d01bf45-7cf8-44f1-a306-8ed97d937dfe-kube-api-access-w5bgs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d01bf45-7cf8-44f1-a306-8ed97d937dfe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.784888 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d01bf45-7cf8-44f1-a306-8ed97d937dfe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 15:37:30 crc kubenswrapper[4789]: I1008 15:37:30.886963 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 15:37:31 crc kubenswrapper[4789]: I1008 15:37:31.328494 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 15:37:31 crc kubenswrapper[4789]: W1008 15:37:31.331140 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d01bf45_7cf8_44f1_a306_8ed97d937dfe.slice/crio-557fa1218e4b161fa8d9d60957d09ac0dd024914bbab9fe80d51bae18e0318b3 WatchSource:0}: Error finding container 557fa1218e4b161fa8d9d60957d09ac0dd024914bbab9fe80d51bae18e0318b3: Status 404 returned error can't find the container with id 557fa1218e4b161fa8d9d60957d09ac0dd024914bbab9fe80d51bae18e0318b3 Oct 08 15:37:31 crc kubenswrapper[4789]: I1008 15:37:31.332777 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:37:32 crc kubenswrapper[4789]: I1008 15:37:32.049799 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"3d01bf45-7cf8-44f1-a306-8ed97d937dfe","Type":"ContainerStarted","Data":"557fa1218e4b161fa8d9d60957d09ac0dd024914bbab9fe80d51bae18e0318b3"} Oct 08 15:37:33 crc kubenswrapper[4789]: I1008 15:37:33.061240 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"3d01bf45-7cf8-44f1-a306-8ed97d937dfe","Type":"ContainerStarted","Data":"be60d7a3082a68271a45c6e2dfb55d6fdbb2029004f78286db6467d347e23b91"} Oct 08 15:37:33 crc kubenswrapper[4789]: I1008 15:37:33.089598 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.292274904 podStartE2EDuration="3.089578594s" podCreationTimestamp="2025-10-08 15:37:30 +0000 UTC" firstStartedPulling="2025-10-08 15:37:31.33259755 +0000 UTC m=+5791.239345042" lastFinishedPulling="2025-10-08 15:37:32.12990124 +0000 UTC m=+5792.036648732" observedRunningTime="2025-10-08 15:37:33.078345818 +0000 UTC m=+5792.985093310" watchObservedRunningTime="2025-10-08 15:37:33.089578594 +0000 UTC m=+5792.996326086" Oct 08 15:37:33 crc kubenswrapper[4789]: I1008 15:37:33.729972 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:37:33 crc kubenswrapper[4789]: E1008 15:37:33.730519 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:37:45 crc kubenswrapper[4789]: I1008 15:37:45.730609 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:37:45 crc kubenswrapper[4789]: E1008 15:37:45.731453 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.558330 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gqlwf/must-gather-hq4p2"] Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.560660 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.562625 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gqlwf"/"kube-root-ca.crt" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.562867 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gqlwf"/"default-dockercfg-2fvlw" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.563722 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gqlwf"/"openshift-service-ca.crt" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.585210 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gqlwf/must-gather-hq4p2"] Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.639286 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmhrl\" (UniqueName: \"kubernetes.io/projected/12cb06f8-6eb0-4468-ac4b-257d2e36b765-kube-api-access-kmhrl\") pod \"must-gather-hq4p2\" (UID: \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\") " pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.640004 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12cb06f8-6eb0-4468-ac4b-257d2e36b765-must-gather-output\") pod \"must-gather-hq4p2\" (UID: \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\") " pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.741908 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12cb06f8-6eb0-4468-ac4b-257d2e36b765-must-gather-output\") pod \"must-gather-hq4p2\" (UID: \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\") " pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.742054 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmhrl\" (UniqueName: \"kubernetes.io/projected/12cb06f8-6eb0-4468-ac4b-257d2e36b765-kube-api-access-kmhrl\") pod \"must-gather-hq4p2\" (UID: \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\") " pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.742554 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12cb06f8-6eb0-4468-ac4b-257d2e36b765-must-gather-output\") pod \"must-gather-hq4p2\" (UID: \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\") " pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.764595 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmhrl\" (UniqueName: \"kubernetes.io/projected/12cb06f8-6eb0-4468-ac4b-257d2e36b765-kube-api-access-kmhrl\") pod \"must-gather-hq4p2\" (UID: \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\") " pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:37:52 crc kubenswrapper[4789]: I1008 15:37:52.879283 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:37:53 crc kubenswrapper[4789]: I1008 15:37:53.366880 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gqlwf/must-gather-hq4p2"] Oct 08 15:37:54 crc kubenswrapper[4789]: I1008 15:37:54.252357 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" event={"ID":"12cb06f8-6eb0-4468-ac4b-257d2e36b765","Type":"ContainerStarted","Data":"9439d61fc0af2645aaec80eabfab5dae95e5c386d9a7a65321a7aa0c529591ae"} Oct 08 15:37:58 crc kubenswrapper[4789]: I1008 15:37:58.732514 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:37:58 crc kubenswrapper[4789]: E1008 15:37:58.733227 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:38:00 crc kubenswrapper[4789]: I1008 15:38:00.319850 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" event={"ID":"12cb06f8-6eb0-4468-ac4b-257d2e36b765","Type":"ContainerStarted","Data":"d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772"} Oct 08 15:38:00 crc kubenswrapper[4789]: I1008 15:38:00.320479 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" event={"ID":"12cb06f8-6eb0-4468-ac4b-257d2e36b765","Type":"ContainerStarted","Data":"97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8"} Oct 08 15:38:00 crc kubenswrapper[4789]: I1008 15:38:00.336432 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" podStartSLOduration=2.290312684 podStartE2EDuration="8.336415129s" podCreationTimestamp="2025-10-08 15:37:52 +0000 UTC" firstStartedPulling="2025-10-08 15:37:53.35811953 +0000 UTC m=+5813.264867022" lastFinishedPulling="2025-10-08 15:37:59.404221975 +0000 UTC m=+5819.310969467" observedRunningTime="2025-10-08 15:38:00.332187984 +0000 UTC m=+5820.238935476" watchObservedRunningTime="2025-10-08 15:38:00.336415129 +0000 UTC m=+5820.243162621" Oct 08 15:38:03 crc kubenswrapper[4789]: I1008 15:38:03.847921 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gqlwf/crc-debug-6x54f"] Oct 08 15:38:03 crc kubenswrapper[4789]: I1008 15:38:03.850878 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:38:03 crc kubenswrapper[4789]: I1008 15:38:03.978556 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt5g2\" (UniqueName: \"kubernetes.io/projected/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-kube-api-access-qt5g2\") pod \"crc-debug-6x54f\" (UID: \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\") " pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:38:03 crc kubenswrapper[4789]: I1008 15:38:03.978861 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-host\") pod \"crc-debug-6x54f\" (UID: \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\") " pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:38:04 crc kubenswrapper[4789]: I1008 15:38:04.080336 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt5g2\" (UniqueName: \"kubernetes.io/projected/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-kube-api-access-qt5g2\") pod \"crc-debug-6x54f\" (UID: \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\") " pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:38:04 crc kubenswrapper[4789]: I1008 15:38:04.080938 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-host\") pod \"crc-debug-6x54f\" (UID: \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\") " pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:38:04 crc kubenswrapper[4789]: I1008 15:38:04.081030 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-host\") pod \"crc-debug-6x54f\" (UID: \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\") " pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:38:04 crc kubenswrapper[4789]: I1008 15:38:04.114321 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt5g2\" (UniqueName: \"kubernetes.io/projected/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-kube-api-access-qt5g2\") pod \"crc-debug-6x54f\" (UID: \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\") " pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:38:04 crc kubenswrapper[4789]: I1008 15:38:04.172115 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:38:04 crc kubenswrapper[4789]: W1008 15:38:04.206142 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b15eab7_790c_40e8_b83a_9c2e9a4bd8cf.slice/crio-77ed05b85942d56c86d559888b8ca25cac65263bbc365a6274c332d15428d414 WatchSource:0}: Error finding container 77ed05b85942d56c86d559888b8ca25cac65263bbc365a6274c332d15428d414: Status 404 returned error can't find the container with id 77ed05b85942d56c86d559888b8ca25cac65263bbc365a6274c332d15428d414 Oct 08 15:38:04 crc kubenswrapper[4789]: I1008 15:38:04.359943 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/crc-debug-6x54f" event={"ID":"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf","Type":"ContainerStarted","Data":"77ed05b85942d56c86d559888b8ca25cac65263bbc365a6274c332d15428d414"} Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.288578 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l7k6b"] Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.292214 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.301773 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l7k6b"] Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.363830 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-catalog-content\") pod \"redhat-operators-l7k6b\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.364081 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-utilities\") pod \"redhat-operators-l7k6b\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.364205 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl4gt\" (UniqueName: \"kubernetes.io/projected/d9b76840-432f-4cef-b336-cea2030b2d72-kube-api-access-zl4gt\") pod \"redhat-operators-l7k6b\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.468116 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl4gt\" (UniqueName: \"kubernetes.io/projected/d9b76840-432f-4cef-b336-cea2030b2d72-kube-api-access-zl4gt\") pod \"redhat-operators-l7k6b\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.468410 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-catalog-content\") pod \"redhat-operators-l7k6b\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.468489 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-utilities\") pod \"redhat-operators-l7k6b\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.469096 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-catalog-content\") pod \"redhat-operators-l7k6b\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.469125 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-utilities\") pod \"redhat-operators-l7k6b\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.501215 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl4gt\" (UniqueName: \"kubernetes.io/projected/d9b76840-432f-4cef-b336-cea2030b2d72-kube-api-access-zl4gt\") pod \"redhat-operators-l7k6b\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.644020 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:12 crc kubenswrapper[4789]: I1008 15:38:12.730066 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:38:12 crc kubenswrapper[4789]: E1008 15:38:12.730383 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:38:16 crc kubenswrapper[4789]: I1008 15:38:16.285749 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l7k6b"] Oct 08 15:38:16 crc kubenswrapper[4789]: I1008 15:38:16.477740 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/crc-debug-6x54f" event={"ID":"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf","Type":"ContainerStarted","Data":"8625cabf3cfb56081793dbdc93ef4a921f6f2f59d7a3b54a98b9dc77dca933b3"} Oct 08 15:38:16 crc kubenswrapper[4789]: I1008 15:38:16.479430 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7k6b" event={"ID":"d9b76840-432f-4cef-b336-cea2030b2d72","Type":"ContainerStarted","Data":"d350b35f297fb1fb9b921fac0ac62e3352a5233b33cbb0a9581c4e2a2174e01e"} Oct 08 15:38:16 crc kubenswrapper[4789]: I1008 15:38:16.519679 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gqlwf/crc-debug-6x54f" podStartSLOduration=1.920822553 podStartE2EDuration="13.51965498s" podCreationTimestamp="2025-10-08 15:38:03 +0000 UTC" firstStartedPulling="2025-10-08 15:38:04.208130057 +0000 UTC m=+5824.114877549" lastFinishedPulling="2025-10-08 15:38:15.806962484 +0000 UTC m=+5835.713709976" observedRunningTime="2025-10-08 15:38:16.492536121 +0000 UTC m=+5836.399283613" watchObservedRunningTime="2025-10-08 15:38:16.51965498 +0000 UTC m=+5836.426402482" Oct 08 15:38:17 crc kubenswrapper[4789]: I1008 15:38:17.490525 4789 generic.go:334] "Generic (PLEG): container finished" podID="d9b76840-432f-4cef-b336-cea2030b2d72" containerID="d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84" exitCode=0 Oct 08 15:38:17 crc kubenswrapper[4789]: I1008 15:38:17.490611 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7k6b" event={"ID":"d9b76840-432f-4cef-b336-cea2030b2d72","Type":"ContainerDied","Data":"d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84"} Oct 08 15:38:18 crc kubenswrapper[4789]: I1008 15:38:18.503951 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7k6b" event={"ID":"d9b76840-432f-4cef-b336-cea2030b2d72","Type":"ContainerStarted","Data":"613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f"} Oct 08 15:38:25 crc kubenswrapper[4789]: I1008 15:38:25.729782 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:38:25 crc kubenswrapper[4789]: E1008 15:38:25.730655 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:38:27 crc kubenswrapper[4789]: I1008 15:38:27.602503 4789 generic.go:334] "Generic (PLEG): container finished" podID="d9b76840-432f-4cef-b336-cea2030b2d72" containerID="613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f" exitCode=0 Oct 08 15:38:27 crc kubenswrapper[4789]: I1008 15:38:27.602567 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7k6b" event={"ID":"d9b76840-432f-4cef-b336-cea2030b2d72","Type":"ContainerDied","Data":"613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f"} Oct 08 15:38:30 crc kubenswrapper[4789]: I1008 15:38:30.637977 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7k6b" event={"ID":"d9b76840-432f-4cef-b336-cea2030b2d72","Type":"ContainerStarted","Data":"9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d"} Oct 08 15:38:30 crc kubenswrapper[4789]: I1008 15:38:30.672861 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l7k6b" podStartSLOduration=8.102492937 podStartE2EDuration="18.672834194s" podCreationTimestamp="2025-10-08 15:38:12 +0000 UTC" firstStartedPulling="2025-10-08 15:38:17.492125773 +0000 UTC m=+5837.398873265" lastFinishedPulling="2025-10-08 15:38:28.06246703 +0000 UTC m=+5847.969214522" observedRunningTime="2025-10-08 15:38:30.663542561 +0000 UTC m=+5850.570290063" watchObservedRunningTime="2025-10-08 15:38:30.672834194 +0000 UTC m=+5850.579581686" Oct 08 15:38:32 crc kubenswrapper[4789]: I1008 15:38:32.644689 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:32 crc kubenswrapper[4789]: I1008 15:38:32.646688 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:33 crc kubenswrapper[4789]: I1008 15:38:33.702563 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l7k6b" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" containerName="registry-server" probeResult="failure" output=< Oct 08 15:38:33 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 15:38:33 crc kubenswrapper[4789]: > Oct 08 15:38:38 crc kubenswrapper[4789]: I1008 15:38:38.730698 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:38:38 crc kubenswrapper[4789]: E1008 15:38:38.731479 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:38:42 crc kubenswrapper[4789]: I1008 15:38:42.706420 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:42 crc kubenswrapper[4789]: I1008 15:38:42.781121 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:43 crc kubenswrapper[4789]: I1008 15:38:43.491854 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l7k6b"] Oct 08 15:38:43 crc kubenswrapper[4789]: I1008 15:38:43.779632 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l7k6b" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" containerName="registry-server" containerID="cri-o://9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d" gracePeriod=2 Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.345981 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.403517 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-utilities\") pod \"d9b76840-432f-4cef-b336-cea2030b2d72\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.403917 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-catalog-content\") pod \"d9b76840-432f-4cef-b336-cea2030b2d72\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.404069 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl4gt\" (UniqueName: \"kubernetes.io/projected/d9b76840-432f-4cef-b336-cea2030b2d72-kube-api-access-zl4gt\") pod \"d9b76840-432f-4cef-b336-cea2030b2d72\" (UID: \"d9b76840-432f-4cef-b336-cea2030b2d72\") " Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.404618 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-utilities" (OuterVolumeSpecName: "utilities") pod "d9b76840-432f-4cef-b336-cea2030b2d72" (UID: "d9b76840-432f-4cef-b336-cea2030b2d72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.404780 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.413879 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b76840-432f-4cef-b336-cea2030b2d72-kube-api-access-zl4gt" (OuterVolumeSpecName: "kube-api-access-zl4gt") pod "d9b76840-432f-4cef-b336-cea2030b2d72" (UID: "d9b76840-432f-4cef-b336-cea2030b2d72"). InnerVolumeSpecName "kube-api-access-zl4gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.490653 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9b76840-432f-4cef-b336-cea2030b2d72" (UID: "d9b76840-432f-4cef-b336-cea2030b2d72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.507567 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl4gt\" (UniqueName: \"kubernetes.io/projected/d9b76840-432f-4cef-b336-cea2030b2d72-kube-api-access-zl4gt\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.507611 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9b76840-432f-4cef-b336-cea2030b2d72-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.791235 4789 generic.go:334] "Generic (PLEG): container finished" podID="d9b76840-432f-4cef-b336-cea2030b2d72" containerID="9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d" exitCode=0 Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.791293 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7k6b" event={"ID":"d9b76840-432f-4cef-b336-cea2030b2d72","Type":"ContainerDied","Data":"9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d"} Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.791336 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7k6b" event={"ID":"d9b76840-432f-4cef-b336-cea2030b2d72","Type":"ContainerDied","Data":"d350b35f297fb1fb9b921fac0ac62e3352a5233b33cbb0a9581c4e2a2174e01e"} Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.791337 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l7k6b" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.791360 4789 scope.go:117] "RemoveContainer" containerID="9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.816500 4789 scope.go:117] "RemoveContainer" containerID="613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.867624 4789 scope.go:117] "RemoveContainer" containerID="d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84" Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.872058 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l7k6b"] Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.883472 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l7k6b"] Oct 08 15:38:44 crc kubenswrapper[4789]: I1008 15:38:44.997397 4789 scope.go:117] "RemoveContainer" containerID="9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d" Oct 08 15:38:45 crc kubenswrapper[4789]: E1008 15:38:45.008216 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d\": container with ID starting with 9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d not found: ID does not exist" containerID="9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d" Oct 08 15:38:45 crc kubenswrapper[4789]: I1008 15:38:45.008282 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d"} err="failed to get container status \"9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d\": rpc error: code = NotFound desc = could not find container \"9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d\": container with ID starting with 9a990754b8fdcdb17081f11617730fbc5e15d898c23cd5f54c2374641f26d20d not found: ID does not exist" Oct 08 15:38:45 crc kubenswrapper[4789]: I1008 15:38:45.008317 4789 scope.go:117] "RemoveContainer" containerID="613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f" Oct 08 15:38:45 crc kubenswrapper[4789]: E1008 15:38:45.012189 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f\": container with ID starting with 613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f not found: ID does not exist" containerID="613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f" Oct 08 15:38:45 crc kubenswrapper[4789]: I1008 15:38:45.012245 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f"} err="failed to get container status \"613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f\": rpc error: code = NotFound desc = could not find container \"613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f\": container with ID starting with 613a53d22049cb9d282ee006a0aeaba5736907c2a671c3ddf5af27223469f40f not found: ID does not exist" Oct 08 15:38:45 crc kubenswrapper[4789]: I1008 15:38:45.012280 4789 scope.go:117] "RemoveContainer" containerID="d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84" Oct 08 15:38:45 crc kubenswrapper[4789]: E1008 15:38:45.012904 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84\": container with ID starting with d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84 not found: ID does not exist" containerID="d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84" Oct 08 15:38:45 crc kubenswrapper[4789]: I1008 15:38:45.012946 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84"} err="failed to get container status \"d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84\": rpc error: code = NotFound desc = could not find container \"d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84\": container with ID starting with d61cac99e2e5973b53df7256d1ffa04b47ed79bacdf25853aac7ba5d7f684c84 not found: ID does not exist" Oct 08 15:38:46 crc kubenswrapper[4789]: I1008 15:38:46.740351 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" path="/var/lib/kubelet/pods/d9b76840-432f-4cef-b336-cea2030b2d72/volumes" Oct 08 15:38:52 crc kubenswrapper[4789]: I1008 15:38:52.730528 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:38:52 crc kubenswrapper[4789]: E1008 15:38:52.731669 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:39:07 crc kubenswrapper[4789]: I1008 15:39:07.730260 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:39:07 crc kubenswrapper[4789]: E1008 15:39:07.731183 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:39:19 crc kubenswrapper[4789]: I1008 15:39:19.729799 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:39:19 crc kubenswrapper[4789]: E1008 15:39:19.730559 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:39:28 crc kubenswrapper[4789]: I1008 15:39:28.542894 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-59fddc7b48-jvk6w_1033617f-6aaf-4d50-bd32-fb0eaf6c99b8/barbican-api/0.log" Oct 08 15:39:28 crc kubenswrapper[4789]: I1008 15:39:28.550860 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-59fddc7b48-jvk6w_1033617f-6aaf-4d50-bd32-fb0eaf6c99b8/barbican-api-log/0.log" Oct 08 15:39:28 crc kubenswrapper[4789]: I1008 15:39:28.756800 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5db9877796-wf2x5_01d2c41c-c841-46c8-84a5-d03987977dba/barbican-keystone-listener/0.log" Oct 08 15:39:28 crc kubenswrapper[4789]: I1008 15:39:28.817540 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5db9877796-wf2x5_01d2c41c-c841-46c8-84a5-d03987977dba/barbican-keystone-listener-log/0.log" Oct 08 15:39:28 crc kubenswrapper[4789]: I1008 15:39:28.983439 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-56d947cc6f-sgvkl_3028d0aa-cb65-4b14-aebb-7b9cb4910eed/barbican-worker/0.log" Oct 08 15:39:29 crc kubenswrapper[4789]: I1008 15:39:29.017885 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-56d947cc6f-sgvkl_3028d0aa-cb65-4b14-aebb-7b9cb4910eed/barbican-worker-log/0.log" Oct 08 15:39:29 crc kubenswrapper[4789]: I1008 15:39:29.231899 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj_5f34622a-a071-45a1-b7fd-75fa8fbb70fe/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:29 crc kubenswrapper[4789]: I1008 15:39:29.533022 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3152634e-3d68-4544-ad93-2999a2273f3a/proxy-httpd/0.log" Oct 08 15:39:29 crc kubenswrapper[4789]: I1008 15:39:29.539443 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3152634e-3d68-4544-ad93-2999a2273f3a/ceilometer-central-agent/0.log" Oct 08 15:39:29 crc kubenswrapper[4789]: I1008 15:39:29.593340 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3152634e-3d68-4544-ad93-2999a2273f3a/ceilometer-notification-agent/0.log" Oct 08 15:39:29 crc kubenswrapper[4789]: I1008 15:39:29.720465 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3152634e-3d68-4544-ad93-2999a2273f3a/sg-core/0.log" Oct 08 15:39:30 crc kubenswrapper[4789]: I1008 15:39:30.001369 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a72d40d2-86d8-406f-9f5c-adec376cc947/cinder-api-log/0.log" Oct 08 15:39:30 crc kubenswrapper[4789]: I1008 15:39:30.331846 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a72d40d2-86d8-406f-9f5c-adec376cc947/cinder-api/0.log" Oct 08 15:39:30 crc kubenswrapper[4789]: I1008 15:39:30.466142 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_6b3acb2d-0c53-40a4-8857-8e07cf54c530/probe/0.log" Oct 08 15:39:30 crc kubenswrapper[4789]: I1008 15:39:30.492902 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_6b3acb2d-0c53-40a4-8857-8e07cf54c530/cinder-backup/0.log" Oct 08 15:39:30 crc kubenswrapper[4789]: I1008 15:39:30.746461 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4717d784-db56-4840-8b93-9be19efb814b/cinder-scheduler/0.log" Oct 08 15:39:30 crc kubenswrapper[4789]: I1008 15:39:30.767032 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4717d784-db56-4840-8b93-9be19efb814b/probe/0.log" Oct 08 15:39:31 crc kubenswrapper[4789]: I1008 15:39:31.015900 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_7b5501af-92e0-4c4a-badd-05089ae1221b/probe/0.log" Oct 08 15:39:31 crc kubenswrapper[4789]: I1008 15:39:31.127338 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_7b5501af-92e0-4c4a-badd-05089ae1221b/cinder-volume/0.log" Oct 08 15:39:31 crc kubenswrapper[4789]: I1008 15:39:31.370847 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_e0006437-3cf9-437c-a3bd-a12f6ef84aae/probe/0.log" Oct 08 15:39:31 crc kubenswrapper[4789]: I1008 15:39:31.425873 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_e0006437-3cf9-437c-a3bd-a12f6ef84aae/cinder-volume/0.log" Oct 08 15:39:31 crc kubenswrapper[4789]: I1008 15:39:31.615656 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9_48dac93b-b8ea-4016-9a48-36fd0c594bf4/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:31 crc kubenswrapper[4789]: I1008 15:39:31.694137 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd_6f2ef14f-0023-4c8d-8153-5dba30e84c17/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:31 crc kubenswrapper[4789]: I1008 15:39:31.730548 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:39:31 crc kubenswrapper[4789]: E1008 15:39:31.730803 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:39:31 crc kubenswrapper[4789]: I1008 15:39:31.880292 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-znv9w_167cf260-5310-4c48-9911-b5257c1ba256/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:32 crc kubenswrapper[4789]: I1008 15:39:32.073473 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-65bf458dd9-l6k8r_dbf93231-9ffa-4434-a4c5-3bf97ee66e71/init/0.log" Oct 08 15:39:32 crc kubenswrapper[4789]: I1008 15:39:32.311690 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-65bf458dd9-l6k8r_dbf93231-9ffa-4434-a4c5-3bf97ee66e71/init/0.log" Oct 08 15:39:32 crc kubenswrapper[4789]: I1008 15:39:32.428064 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd_617364bf-6612-48bf-bfb4-0eeaf9ff2c60/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:32 crc kubenswrapper[4789]: I1008 15:39:32.505285 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-65bf458dd9-l6k8r_dbf93231-9ffa-4434-a4c5-3bf97ee66e71/dnsmasq-dns/0.log" Oct 08 15:39:32 crc kubenswrapper[4789]: I1008 15:39:32.582304 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1fbecfec-ecd2-458e-b04f-e715a9ec4186/glance-httpd/0.log" Oct 08 15:39:32 crc kubenswrapper[4789]: I1008 15:39:32.615669 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1fbecfec-ecd2-458e-b04f-e715a9ec4186/glance-log/0.log" Oct 08 15:39:32 crc kubenswrapper[4789]: I1008 15:39:32.800747 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a8fbeb79-b731-40fe-867d-1df6c03db681/glance-log/0.log" Oct 08 15:39:32 crc kubenswrapper[4789]: I1008 15:39:32.841417 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a8fbeb79-b731-40fe-867d-1df6c03db681/glance-httpd/0.log" Oct 08 15:39:33 crc kubenswrapper[4789]: I1008 15:39:33.089281 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6c7bdc9dd-bg9vf_f9289b28-1fb4-47f8-aec3-3d9d616837d3/horizon/0.log" Oct 08 15:39:33 crc kubenswrapper[4789]: I1008 15:39:33.141971 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm_92d5b86d-9984-45c5-8287-7b3094b4332f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:33 crc kubenswrapper[4789]: I1008 15:39:33.444956 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-xpffg_5a8f80d9-3b33-479c-9d6a-77867520218d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:33 crc kubenswrapper[4789]: I1008 15:39:33.682236 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6c7bdc9dd-bg9vf_f9289b28-1fb4-47f8-aec3-3d9d616837d3/horizon-log/0.log" Oct 08 15:39:33 crc kubenswrapper[4789]: I1008 15:39:33.737358 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29332261-6nq76_caa89bd3-cdc8-412a-82f0-1289cf0028c1/keystone-cron/0.log" Oct 08 15:39:33 crc kubenswrapper[4789]: I1008 15:39:33.937957 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_08e907c7-3b60-4d50-977c-32d05a3ce80f/kube-state-metrics/0.log" Oct 08 15:39:34 crc kubenswrapper[4789]: I1008 15:39:34.228218 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg_2066f52b-bb99-4a61-98e4-63f53b8ec7c7/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:34 crc kubenswrapper[4789]: I1008 15:39:34.304262 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c9c77979c-fl8sh_5eeed1f7-b400-4c25-851d-43a1f036fa8b/keystone-api/0.log" Oct 08 15:39:34 crc kubenswrapper[4789]: I1008 15:39:34.899406 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7dcb5d4fcc-q9spz_1532f955-2e73-4a9a-b76a-507c2c545db0/neutron-httpd/0.log" Oct 08 15:39:34 crc kubenswrapper[4789]: I1008 15:39:34.971004 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7dcb5d4fcc-q9spz_1532f955-2e73-4a9a-b76a-507c2c545db0/neutron-api/0.log" Oct 08 15:39:35 crc kubenswrapper[4789]: I1008 15:39:35.079446 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt_7f2ef2f7-aa9f-4767-a032-e461dde85655/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:35 crc kubenswrapper[4789]: I1008 15:39:35.956252 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_de1fa794-a227-446e-a3b0-2b1c120e4ecf/nova-cell0-conductor-conductor/0.log" Oct 08 15:39:36 crc kubenswrapper[4789]: I1008 15:39:36.603428 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_af59a4db-7cb6-4e1f-8381-43e0f13e5875/nova-cell1-conductor-conductor/0.log" Oct 08 15:39:36 crc kubenswrapper[4789]: I1008 15:39:36.857320 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c/nova-api-log/0.log" Oct 08 15:39:37 crc kubenswrapper[4789]: I1008 15:39:37.138897 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c/nova-api-api/0.log" Oct 08 15:39:37 crc kubenswrapper[4789]: I1008 15:39:37.213911 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_42136e0e-1d9e-44ea-b9fc-3e7e313fe107/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 15:39:37 crc kubenswrapper[4789]: I1008 15:39:37.395694 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-gbfjb_b8ce1628-fa93-4910-af5a-ebce3d8ecc30/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:37 crc kubenswrapper[4789]: I1008 15:39:37.536691 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9255b8e1-bd22-43cf-b223-87021f091449/nova-metadata-log/0.log" Oct 08 15:39:38 crc kubenswrapper[4789]: I1008 15:39:38.101640 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ad6bc841-5112-4ce6-a8ae-926e8b7d8043/nova-scheduler-scheduler/0.log" Oct 08 15:39:38 crc kubenswrapper[4789]: I1008 15:39:38.163527 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_32ceb31c-1751-4a04-a91a-b317f3ba3d78/mysql-bootstrap/0.log" Oct 08 15:39:38 crc kubenswrapper[4789]: I1008 15:39:38.422455 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_32ceb31c-1751-4a04-a91a-b317f3ba3d78/mysql-bootstrap/0.log" Oct 08 15:39:38 crc kubenswrapper[4789]: I1008 15:39:38.453327 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_32ceb31c-1751-4a04-a91a-b317f3ba3d78/galera/0.log" Oct 08 15:39:38 crc kubenswrapper[4789]: I1008 15:39:38.672217 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_505cf346-64b1-4de8-8cf6-fca4ddd34cd6/mysql-bootstrap/0.log" Oct 08 15:39:38 crc kubenswrapper[4789]: I1008 15:39:38.931617 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_505cf346-64b1-4de8-8cf6-fca4ddd34cd6/galera/0.log" Oct 08 15:39:38 crc kubenswrapper[4789]: I1008 15:39:38.942563 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_505cf346-64b1-4de8-8cf6-fca4ddd34cd6/mysql-bootstrap/0.log" Oct 08 15:39:39 crc kubenswrapper[4789]: I1008 15:39:39.160794 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2ba79b7a-39cd-4f22-8094-672ef7b58e76/openstackclient/0.log" Oct 08 15:39:39 crc kubenswrapper[4789]: I1008 15:39:39.442375 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nrhvq_23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718/openstack-network-exporter/0.log" Oct 08 15:39:39 crc kubenswrapper[4789]: I1008 15:39:39.733203 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tkv8v_5e2285a3-7296-490c-8649-adeae9a0cabc/ovsdb-server-init/0.log" Oct 08 15:39:39 crc kubenswrapper[4789]: I1008 15:39:39.849150 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9255b8e1-bd22-43cf-b223-87021f091449/nova-metadata-metadata/0.log" Oct 08 15:39:39 crc kubenswrapper[4789]: I1008 15:39:39.948076 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tkv8v_5e2285a3-7296-490c-8649-adeae9a0cabc/ovsdb-server-init/0.log" Oct 08 15:39:40 crc kubenswrapper[4789]: I1008 15:39:40.074740 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tkv8v_5e2285a3-7296-490c-8649-adeae9a0cabc/ovsdb-server/0.log" Oct 08 15:39:40 crc kubenswrapper[4789]: I1008 15:39:40.290337 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zrdhb_91ddb1cc-17c9-4f38-ac8a-d78d10798178/ovn-controller/0.log" Oct 08 15:39:40 crc kubenswrapper[4789]: I1008 15:39:40.327915 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tkv8v_5e2285a3-7296-490c-8649-adeae9a0cabc/ovs-vswitchd/0.log" Oct 08 15:39:40 crc kubenswrapper[4789]: I1008 15:39:40.625432 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4djc2_516d476a-1a73-4e62-a06f-d9eb90541500/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:40 crc kubenswrapper[4789]: I1008 15:39:40.709810 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fd23ae36-37d1-458e-ab17-0f0db57e57d3/openstack-network-exporter/0.log" Oct 08 15:39:40 crc kubenswrapper[4789]: I1008 15:39:40.805439 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fd23ae36-37d1-458e-ab17-0f0db57e57d3/ovn-northd/0.log" Oct 08 15:39:40 crc kubenswrapper[4789]: I1008 15:39:40.967437 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a1a9507e-04fe-4746-a31b-5ad097c48602/openstack-network-exporter/0.log" Oct 08 15:39:41 crc kubenswrapper[4789]: I1008 15:39:41.005964 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a1a9507e-04fe-4746-a31b-5ad097c48602/ovsdbserver-nb/0.log" Oct 08 15:39:41 crc kubenswrapper[4789]: I1008 15:39:41.233604 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_70dc7aa5-b81d-4150-b1dd-57ec365e495c/openstack-network-exporter/0.log" Oct 08 15:39:41 crc kubenswrapper[4789]: I1008 15:39:41.280412 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_70dc7aa5-b81d-4150-b1dd-57ec365e495c/ovsdbserver-sb/0.log" Oct 08 15:39:41 crc kubenswrapper[4789]: I1008 15:39:41.643421 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6c99f9dffd-c2zmb_c79d9b2e-fb14-4d73-a6f0-d8b0684001ac/placement-api/0.log" Oct 08 15:39:41 crc kubenswrapper[4789]: I1008 15:39:41.931959 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/init-config-reloader/0.log" Oct 08 15:39:41 crc kubenswrapper[4789]: I1008 15:39:41.947388 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6c99f9dffd-c2zmb_c79d9b2e-fb14-4d73-a6f0-d8b0684001ac/placement-log/0.log" Oct 08 15:39:42 crc kubenswrapper[4789]: I1008 15:39:42.170371 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/prometheus/0.log" Oct 08 15:39:42 crc kubenswrapper[4789]: I1008 15:39:42.182780 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/config-reloader/0.log" Oct 08 15:39:42 crc kubenswrapper[4789]: I1008 15:39:42.226767 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/init-config-reloader/0.log" Oct 08 15:39:42 crc kubenswrapper[4789]: I1008 15:39:42.398263 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/thanos-sidecar/0.log" Oct 08 15:39:42 crc kubenswrapper[4789]: I1008 15:39:42.454831 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c4758067-9242-48d2-a066-e007669ef85c/setup-container/0.log" Oct 08 15:39:42 crc kubenswrapper[4789]: I1008 15:39:42.664347 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c4758067-9242-48d2-a066-e007669ef85c/setup-container/0.log" Oct 08 15:39:42 crc kubenswrapper[4789]: I1008 15:39:42.737103 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c4758067-9242-48d2-a066-e007669ef85c/rabbitmq/0.log" Oct 08 15:39:42 crc kubenswrapper[4789]: I1008 15:39:42.901981 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_55fc0a2c-48c9-4508-ae9d-5350f3298621/setup-container/0.log" Oct 08 15:39:43 crc kubenswrapper[4789]: I1008 15:39:43.086399 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_55fc0a2c-48c9-4508-ae9d-5350f3298621/setup-container/0.log" Oct 08 15:39:43 crc kubenswrapper[4789]: I1008 15:39:43.142909 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_55fc0a2c-48c9-4508-ae9d-5350f3298621/rabbitmq/0.log" Oct 08 15:39:43 crc kubenswrapper[4789]: I1008 15:39:43.317852 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_698f294c-233d-45d0-a47b-f3580b23b955/setup-container/0.log" Oct 08 15:39:43 crc kubenswrapper[4789]: I1008 15:39:43.538407 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_698f294c-233d-45d0-a47b-f3580b23b955/setup-container/0.log" Oct 08 15:39:43 crc kubenswrapper[4789]: I1008 15:39:43.540768 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_698f294c-233d-45d0-a47b-f3580b23b955/rabbitmq/0.log" Oct 08 15:39:43 crc kubenswrapper[4789]: I1008 15:39:43.729839 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:39:43 crc kubenswrapper[4789]: E1008 15:39:43.730219 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:39:43 crc kubenswrapper[4789]: I1008 15:39:43.757920 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg_f2575ace-ab3d-475e-b049-a2d2614104bf/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:43 crc kubenswrapper[4789]: I1008 15:39:43.882872 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-785jj_c5a284ba-b502-45cc-b9b6-f460908bee2f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:43 crc kubenswrapper[4789]: I1008 15:39:43.990165 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr_484052d0-6aeb-4f53-bad2-d6228aaadcf3/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:44 crc kubenswrapper[4789]: I1008 15:39:44.224857 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-xz5th_1ef78bfb-47eb-40df-93d6-7b6070c33f5f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:44 crc kubenswrapper[4789]: I1008 15:39:44.397660 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-29nmw_ad2db88c-41b1-4feb-b579-96bb456cea2a/ssh-known-hosts-edpm-deployment/0.log" Oct 08 15:39:44 crc kubenswrapper[4789]: I1008 15:39:44.659296 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7bf8478fbc-fpmxn_ba045b83-d6ec-4760-98dd-b330dca8ed99/proxy-server/0.log" Oct 08 15:39:44 crc kubenswrapper[4789]: I1008 15:39:44.779614 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7bf8478fbc-fpmxn_ba045b83-d6ec-4760-98dd-b330dca8ed99/proxy-httpd/0.log" Oct 08 15:39:44 crc kubenswrapper[4789]: I1008 15:39:44.894439 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-pr6sr_a7b17c27-5518-480a-875f-091ae1dd3606/swift-ring-rebalance/0.log" Oct 08 15:39:44 crc kubenswrapper[4789]: I1008 15:39:44.992532 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/account-auditor/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.146832 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/account-reaper/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.301874 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/account-server/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.319248 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/account-replicator/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.349334 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/container-auditor/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.555504 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/container-replicator/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.588404 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/container-server/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.591425 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/container-updater/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.783470 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-expirer/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.792378 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-auditor/0.log" Oct 08 15:39:45 crc kubenswrapper[4789]: I1008 15:39:45.860973 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-replicator/0.log" Oct 08 15:39:46 crc kubenswrapper[4789]: I1008 15:39:46.057926 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-updater/0.log" Oct 08 15:39:46 crc kubenswrapper[4789]: I1008 15:39:46.094149 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-server/0.log" Oct 08 15:39:46 crc kubenswrapper[4789]: I1008 15:39:46.143381 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/rsync/0.log" Oct 08 15:39:46 crc kubenswrapper[4789]: I1008 15:39:46.298614 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/swift-recon-cron/0.log" Oct 08 15:39:46 crc kubenswrapper[4789]: I1008 15:39:46.385618 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk_c9031f68-f885-456b-a1ea-ce7742e0eccc/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:46 crc kubenswrapper[4789]: I1008 15:39:46.685258 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5622aa64-be18-4420-8c84-b76fe03c1346/tempest-tests-tempest-tests-runner/0.log" Oct 08 15:39:46 crc kubenswrapper[4789]: I1008 15:39:46.753422 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_3d01bf45-7cf8-44f1-a306-8ed97d937dfe/test-operator-logs-container/0.log" Oct 08 15:39:46 crc kubenswrapper[4789]: I1008 15:39:46.953720 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj_04be0498-514b-45df-9ff0-7eae07586037/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:39:48 crc kubenswrapper[4789]: I1008 15:39:48.410640 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_d84335e3-9c9e-473f-be0a-f1f4535ae6b5/watcher-applier/0.log" Oct 08 15:39:48 crc kubenswrapper[4789]: I1008 15:39:48.804057 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_4bdbb9f8-14b6-4f34-b818-a9deb84247cf/watcher-api-log/0.log" Oct 08 15:39:52 crc kubenswrapper[4789]: I1008 15:39:52.607135 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_4c9e2e8c-9df6-4340-af61-2778209a54dd/watcher-decision-engine/0.log" Oct 08 15:39:53 crc kubenswrapper[4789]: I1008 15:39:53.260746 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_4bdbb9f8-14b6-4f34-b818-a9deb84247cf/watcher-api/0.log" Oct 08 15:39:55 crc kubenswrapper[4789]: I1008 15:39:55.730225 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:39:55 crc kubenswrapper[4789]: E1008 15:39:55.730775 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:40:04 crc kubenswrapper[4789]: I1008 15:40:04.124954 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_78e021d8-ac60-4092-9e49-bbb3fff88aa7/memcached/0.log" Oct 08 15:40:10 crc kubenswrapper[4789]: I1008 15:40:10.737138 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:40:10 crc kubenswrapper[4789]: E1008 15:40:10.737955 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:40:25 crc kubenswrapper[4789]: I1008 15:40:25.730036 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:40:25 crc kubenswrapper[4789]: E1008 15:40:25.731002 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:40:35 crc kubenswrapper[4789]: I1008 15:40:35.945140 4789 generic.go:334] "Generic (PLEG): container finished" podID="7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf" containerID="8625cabf3cfb56081793dbdc93ef4a921f6f2f59d7a3b54a98b9dc77dca933b3" exitCode=0 Oct 08 15:40:35 crc kubenswrapper[4789]: I1008 15:40:35.945200 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/crc-debug-6x54f" event={"ID":"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf","Type":"ContainerDied","Data":"8625cabf3cfb56081793dbdc93ef4a921f6f2f59d7a3b54a98b9dc77dca933b3"} Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.067615 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.101056 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gqlwf/crc-debug-6x54f"] Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.109412 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gqlwf/crc-debug-6x54f"] Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.230822 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt5g2\" (UniqueName: \"kubernetes.io/projected/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-kube-api-access-qt5g2\") pod \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\" (UID: \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\") " Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.231306 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-host\") pod \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\" (UID: \"7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf\") " Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.231341 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-host" (OuterVolumeSpecName: "host") pod "7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf" (UID: "7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.231847 4789 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-host\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.235668 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-kube-api-access-qt5g2" (OuterVolumeSpecName: "kube-api-access-qt5g2") pod "7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf" (UID: "7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf"). InnerVolumeSpecName "kube-api-access-qt5g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.333801 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt5g2\" (UniqueName: \"kubernetes.io/projected/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf-kube-api-access-qt5g2\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.730956 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:40:37 crc kubenswrapper[4789]: E1008 15:40:37.731307 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.964564 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77ed05b85942d56c86d559888b8ca25cac65263bbc365a6274c332d15428d414" Oct 08 15:40:37 crc kubenswrapper[4789]: I1008 15:40:37.964615 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-6x54f" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.282247 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gqlwf/crc-debug-qwlkb"] Oct 08 15:40:38 crc kubenswrapper[4789]: E1008 15:40:38.282773 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" containerName="extract-content" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.283099 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" containerName="extract-content" Oct 08 15:40:38 crc kubenswrapper[4789]: E1008 15:40:38.283145 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf" containerName="container-00" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.283154 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf" containerName="container-00" Oct 08 15:40:38 crc kubenswrapper[4789]: E1008 15:40:38.283183 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" containerName="registry-server" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.283191 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" containerName="registry-server" Oct 08 15:40:38 crc kubenswrapper[4789]: E1008 15:40:38.283204 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" containerName="extract-utilities" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.283211 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" containerName="extract-utilities" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.283456 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf" containerName="container-00" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.283479 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b76840-432f-4cef-b336-cea2030b2d72" containerName="registry-server" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.284403 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.456437 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d17b4df-3206-421b-be6f-b25b237364e4-host\") pod \"crc-debug-qwlkb\" (UID: \"6d17b4df-3206-421b-be6f-b25b237364e4\") " pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.456544 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spvkm\" (UniqueName: \"kubernetes.io/projected/6d17b4df-3206-421b-be6f-b25b237364e4-kube-api-access-spvkm\") pod \"crc-debug-qwlkb\" (UID: \"6d17b4df-3206-421b-be6f-b25b237364e4\") " pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.558213 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d17b4df-3206-421b-be6f-b25b237364e4-host\") pod \"crc-debug-qwlkb\" (UID: \"6d17b4df-3206-421b-be6f-b25b237364e4\") " pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.558287 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spvkm\" (UniqueName: \"kubernetes.io/projected/6d17b4df-3206-421b-be6f-b25b237364e4-kube-api-access-spvkm\") pod \"crc-debug-qwlkb\" (UID: \"6d17b4df-3206-421b-be6f-b25b237364e4\") " pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.558360 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d17b4df-3206-421b-be6f-b25b237364e4-host\") pod \"crc-debug-qwlkb\" (UID: \"6d17b4df-3206-421b-be6f-b25b237364e4\") " pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.577098 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spvkm\" (UniqueName: \"kubernetes.io/projected/6d17b4df-3206-421b-be6f-b25b237364e4-kube-api-access-spvkm\") pod \"crc-debug-qwlkb\" (UID: \"6d17b4df-3206-421b-be6f-b25b237364e4\") " pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.604651 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.745339 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf" path="/var/lib/kubelet/pods/7b15eab7-790c-40e8-b83a-9c2e9a4bd8cf/volumes" Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.976512 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" event={"ID":"6d17b4df-3206-421b-be6f-b25b237364e4","Type":"ContainerStarted","Data":"36b4f9eece173562ab3740c21d7afdcdc43d1bcb1de8b6a728eb2cb99185a819"} Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.976555 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" event={"ID":"6d17b4df-3206-421b-be6f-b25b237364e4","Type":"ContainerStarted","Data":"c39e3bfb11c33f1a3bc1fd7c2a51fafeb9da813abf3908f78603f561e71e4efb"} Oct 08 15:40:38 crc kubenswrapper[4789]: I1008 15:40:38.992267 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" podStartSLOduration=0.992243059 podStartE2EDuration="992.243059ms" podCreationTimestamp="2025-10-08 15:40:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:40:38.989238017 +0000 UTC m=+5978.895985519" watchObservedRunningTime="2025-10-08 15:40:38.992243059 +0000 UTC m=+5978.898990551" Oct 08 15:40:39 crc kubenswrapper[4789]: I1008 15:40:39.991439 4789 generic.go:334] "Generic (PLEG): container finished" podID="6d17b4df-3206-421b-be6f-b25b237364e4" containerID="36b4f9eece173562ab3740c21d7afdcdc43d1bcb1de8b6a728eb2cb99185a819" exitCode=0 Oct 08 15:40:39 crc kubenswrapper[4789]: I1008 15:40:39.991547 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" event={"ID":"6d17b4df-3206-421b-be6f-b25b237364e4","Type":"ContainerDied","Data":"36b4f9eece173562ab3740c21d7afdcdc43d1bcb1de8b6a728eb2cb99185a819"} Oct 08 15:40:41 crc kubenswrapper[4789]: I1008 15:40:41.129042 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:41 crc kubenswrapper[4789]: I1008 15:40:41.315013 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spvkm\" (UniqueName: \"kubernetes.io/projected/6d17b4df-3206-421b-be6f-b25b237364e4-kube-api-access-spvkm\") pod \"6d17b4df-3206-421b-be6f-b25b237364e4\" (UID: \"6d17b4df-3206-421b-be6f-b25b237364e4\") " Oct 08 15:40:41 crc kubenswrapper[4789]: I1008 15:40:41.315110 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d17b4df-3206-421b-be6f-b25b237364e4-host\") pod \"6d17b4df-3206-421b-be6f-b25b237364e4\" (UID: \"6d17b4df-3206-421b-be6f-b25b237364e4\") " Oct 08 15:40:41 crc kubenswrapper[4789]: I1008 15:40:41.315299 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d17b4df-3206-421b-be6f-b25b237364e4-host" (OuterVolumeSpecName: "host") pod "6d17b4df-3206-421b-be6f-b25b237364e4" (UID: "6d17b4df-3206-421b-be6f-b25b237364e4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:40:41 crc kubenswrapper[4789]: I1008 15:40:41.315658 4789 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d17b4df-3206-421b-be6f-b25b237364e4-host\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:41 crc kubenswrapper[4789]: I1008 15:40:41.320438 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d17b4df-3206-421b-be6f-b25b237364e4-kube-api-access-spvkm" (OuterVolumeSpecName: "kube-api-access-spvkm") pod "6d17b4df-3206-421b-be6f-b25b237364e4" (UID: "6d17b4df-3206-421b-be6f-b25b237364e4"). InnerVolumeSpecName "kube-api-access-spvkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:40:41 crc kubenswrapper[4789]: I1008 15:40:41.417429 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spvkm\" (UniqueName: \"kubernetes.io/projected/6d17b4df-3206-421b-be6f-b25b237364e4-kube-api-access-spvkm\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:42 crc kubenswrapper[4789]: I1008 15:40:42.010782 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" event={"ID":"6d17b4df-3206-421b-be6f-b25b237364e4","Type":"ContainerDied","Data":"c39e3bfb11c33f1a3bc1fd7c2a51fafeb9da813abf3908f78603f561e71e4efb"} Oct 08 15:40:42 crc kubenswrapper[4789]: I1008 15:40:42.010816 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c39e3bfb11c33f1a3bc1fd7c2a51fafeb9da813abf3908f78603f561e71e4efb" Oct 08 15:40:42 crc kubenswrapper[4789]: I1008 15:40:42.010873 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-qwlkb" Oct 08 15:40:48 crc kubenswrapper[4789]: I1008 15:40:48.729775 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:40:48 crc kubenswrapper[4789]: E1008 15:40:48.730583 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:40:48 crc kubenswrapper[4789]: I1008 15:40:48.765709 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gqlwf/crc-debug-qwlkb"] Oct 08 15:40:48 crc kubenswrapper[4789]: I1008 15:40:48.774216 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gqlwf/crc-debug-qwlkb"] Oct 08 15:40:49 crc kubenswrapper[4789]: I1008 15:40:49.928845 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gqlwf/crc-debug-vjkkb"] Oct 08 15:40:49 crc kubenswrapper[4789]: E1008 15:40:49.929668 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d17b4df-3206-421b-be6f-b25b237364e4" containerName="container-00" Oct 08 15:40:49 crc kubenswrapper[4789]: I1008 15:40:49.929687 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d17b4df-3206-421b-be6f-b25b237364e4" containerName="container-00" Oct 08 15:40:49 crc kubenswrapper[4789]: I1008 15:40:49.929979 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d17b4df-3206-421b-be6f-b25b237364e4" containerName="container-00" Oct 08 15:40:49 crc kubenswrapper[4789]: I1008 15:40:49.930822 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:50 crc kubenswrapper[4789]: I1008 15:40:50.086156 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/334a1212-ea46-4ef4-8770-93ce8c09dfce-host\") pod \"crc-debug-vjkkb\" (UID: \"334a1212-ea46-4ef4-8770-93ce8c09dfce\") " pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:50 crc kubenswrapper[4789]: I1008 15:40:50.086465 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xtpj\" (UniqueName: \"kubernetes.io/projected/334a1212-ea46-4ef4-8770-93ce8c09dfce-kube-api-access-9xtpj\") pod \"crc-debug-vjkkb\" (UID: \"334a1212-ea46-4ef4-8770-93ce8c09dfce\") " pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:50 crc kubenswrapper[4789]: I1008 15:40:50.188592 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/334a1212-ea46-4ef4-8770-93ce8c09dfce-host\") pod \"crc-debug-vjkkb\" (UID: \"334a1212-ea46-4ef4-8770-93ce8c09dfce\") " pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:50 crc kubenswrapper[4789]: I1008 15:40:50.188749 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/334a1212-ea46-4ef4-8770-93ce8c09dfce-host\") pod \"crc-debug-vjkkb\" (UID: \"334a1212-ea46-4ef4-8770-93ce8c09dfce\") " pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:50 crc kubenswrapper[4789]: I1008 15:40:50.188774 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xtpj\" (UniqueName: \"kubernetes.io/projected/334a1212-ea46-4ef4-8770-93ce8c09dfce-kube-api-access-9xtpj\") pod \"crc-debug-vjkkb\" (UID: \"334a1212-ea46-4ef4-8770-93ce8c09dfce\") " pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:50 crc kubenswrapper[4789]: I1008 15:40:50.209486 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xtpj\" (UniqueName: \"kubernetes.io/projected/334a1212-ea46-4ef4-8770-93ce8c09dfce-kube-api-access-9xtpj\") pod \"crc-debug-vjkkb\" (UID: \"334a1212-ea46-4ef4-8770-93ce8c09dfce\") " pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:50 crc kubenswrapper[4789]: I1008 15:40:50.254104 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:50 crc kubenswrapper[4789]: I1008 15:40:50.753769 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d17b4df-3206-421b-be6f-b25b237364e4" path="/var/lib/kubelet/pods/6d17b4df-3206-421b-be6f-b25b237364e4/volumes" Oct 08 15:40:51 crc kubenswrapper[4789]: I1008 15:40:51.101814 4789 generic.go:334] "Generic (PLEG): container finished" podID="334a1212-ea46-4ef4-8770-93ce8c09dfce" containerID="9c6e916129cf69baf0e0d380304aa6bfc948576671eb042638d4e5dc079a8919" exitCode=0 Oct 08 15:40:51 crc kubenswrapper[4789]: I1008 15:40:51.101882 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" event={"ID":"334a1212-ea46-4ef4-8770-93ce8c09dfce","Type":"ContainerDied","Data":"9c6e916129cf69baf0e0d380304aa6bfc948576671eb042638d4e5dc079a8919"} Oct 08 15:40:51 crc kubenswrapper[4789]: I1008 15:40:51.101947 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" event={"ID":"334a1212-ea46-4ef4-8770-93ce8c09dfce","Type":"ContainerStarted","Data":"a1cb87e52fea270852b9e58cf677a3e1ec803122d5da2c19c3ddaea4b5e75743"} Oct 08 15:40:51 crc kubenswrapper[4789]: I1008 15:40:51.149852 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gqlwf/crc-debug-vjkkb"] Oct 08 15:40:51 crc kubenswrapper[4789]: I1008 15:40:51.163183 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gqlwf/crc-debug-vjkkb"] Oct 08 15:40:52 crc kubenswrapper[4789]: I1008 15:40:52.259185 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:52 crc kubenswrapper[4789]: I1008 15:40:52.438790 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xtpj\" (UniqueName: \"kubernetes.io/projected/334a1212-ea46-4ef4-8770-93ce8c09dfce-kube-api-access-9xtpj\") pod \"334a1212-ea46-4ef4-8770-93ce8c09dfce\" (UID: \"334a1212-ea46-4ef4-8770-93ce8c09dfce\") " Oct 08 15:40:52 crc kubenswrapper[4789]: I1008 15:40:52.438854 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/334a1212-ea46-4ef4-8770-93ce8c09dfce-host\") pod \"334a1212-ea46-4ef4-8770-93ce8c09dfce\" (UID: \"334a1212-ea46-4ef4-8770-93ce8c09dfce\") " Oct 08 15:40:52 crc kubenswrapper[4789]: I1008 15:40:52.438932 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/334a1212-ea46-4ef4-8770-93ce8c09dfce-host" (OuterVolumeSpecName: "host") pod "334a1212-ea46-4ef4-8770-93ce8c09dfce" (UID: "334a1212-ea46-4ef4-8770-93ce8c09dfce"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:40:52 crc kubenswrapper[4789]: I1008 15:40:52.439449 4789 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/334a1212-ea46-4ef4-8770-93ce8c09dfce-host\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:52 crc kubenswrapper[4789]: I1008 15:40:52.453205 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/334a1212-ea46-4ef4-8770-93ce8c09dfce-kube-api-access-9xtpj" (OuterVolumeSpecName: "kube-api-access-9xtpj") pod "334a1212-ea46-4ef4-8770-93ce8c09dfce" (UID: "334a1212-ea46-4ef4-8770-93ce8c09dfce"). InnerVolumeSpecName "kube-api-access-9xtpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:40:52 crc kubenswrapper[4789]: I1008 15:40:52.541582 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xtpj\" (UniqueName: \"kubernetes.io/projected/334a1212-ea46-4ef4-8770-93ce8c09dfce-kube-api-access-9xtpj\") on node \"crc\" DevicePath \"\"" Oct 08 15:40:52 crc kubenswrapper[4789]: I1008 15:40:52.742820 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="334a1212-ea46-4ef4-8770-93ce8c09dfce" path="/var/lib/kubelet/pods/334a1212-ea46-4ef4-8770-93ce8c09dfce/volumes" Oct 08 15:40:52 crc kubenswrapper[4789]: I1008 15:40:52.951472 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/util/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.121101 4789 scope.go:117] "RemoveContainer" containerID="9c6e916129cf69baf0e0d380304aa6bfc948576671eb042638d4e5dc079a8919" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.121134 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/crc-debug-vjkkb" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.139580 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/pull/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.178110 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/pull/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.188807 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/util/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.340280 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/util/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.391202 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/extract/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.395209 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/pull/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.529308 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-77hn5_20bc6d13-f85d-423a-9782-097696fe3a0a/kube-rbac-proxy/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.679996 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-77hn5_20bc6d13-f85d-423a-9782-097696fe3a0a/manager/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.687439 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-k8txt_78f0cb36-5bad-4021-b464-c5da9318404a/kube-rbac-proxy/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.791552 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-k8txt_78f0cb36-5bad-4021-b464-c5da9318404a/manager/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.873080 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ptjbw_0aa40208-3985-4820-809f-ca1f94b05036/manager/0.log" Oct 08 15:40:53 crc kubenswrapper[4789]: I1008 15:40:53.883963 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ptjbw_0aa40208-3985-4820-809f-ca1f94b05036/kube-rbac-proxy/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.070081 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-tggcc_00c51d73-7d74-484a-8773-41cbb40c69f9/kube-rbac-proxy/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.158350 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-tggcc_00c51d73-7d74-484a-8773-41cbb40c69f9/manager/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.215059 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-vbtkd_0d8d4170-6f10-4d04-b443-297aada2dc10/kube-rbac-proxy/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.289087 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-vbtkd_0d8d4170-6f10-4d04-b443-297aada2dc10/manager/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.356743 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-w5bgc_f46571c0-32a7-4202-b308-54446adfe3d1/kube-rbac-proxy/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.451970 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-w5bgc_f46571c0-32a7-4202-b308-54446adfe3d1/manager/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.511951 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-wl7d2_e8242a68-738b-4fcd-abaa-22a18790ea28/kube-rbac-proxy/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.748724 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-zslsw_bff33916-174d-4fd4-a5e4-91f7cacb6e52/kube-rbac-proxy/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.754082 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-wl7d2_e8242a68-738b-4fcd-abaa-22a18790ea28/manager/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.758204 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-zslsw_bff33916-174d-4fd4-a5e4-91f7cacb6e52/manager/0.log" Oct 08 15:40:54 crc kubenswrapper[4789]: I1008 15:40:54.914168 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-bvjfl_f7b65ead-d9d5-4f06-80de-ae5c85dcbd45/kube-rbac-proxy/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.041306 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-bvjfl_f7b65ead-d9d5-4f06-80de-ae5c85dcbd45/manager/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.116017 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-pzg7c_8e51126b-8f41-4627-9f68-fb1255bda5bd/kube-rbac-proxy/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.139594 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-pzg7c_8e51126b-8f41-4627-9f68-fb1255bda5bd/manager/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.223289 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-7gsbz_3c0f54fe-d45f-44c2-9f09-4065c9621903/kube-rbac-proxy/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.314866 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-7gsbz_3c0f54fe-d45f-44c2-9f09-4065c9621903/manager/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.401746 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-2j5bd_c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4/kube-rbac-proxy/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.500543 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-2j5bd_c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4/manager/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.522170 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-7trlm_c9e3e3c8-447f-4880-b54b-1d5601271216/kube-rbac-proxy/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.718299 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-2kzff_4e7cd567-305f-417d-bac8-7bf4caf22693/kube-rbac-proxy/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.734485 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-7trlm_c9e3e3c8-447f-4880-b54b-1d5601271216/manager/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.752594 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-2kzff_4e7cd567-305f-417d-bac8-7bf4caf22693/manager/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.891280 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h_6f893fd6-e5cc-42b9-ac46-77567818c0ef/kube-rbac-proxy/0.log" Oct 08 15:40:55 crc kubenswrapper[4789]: I1008 15:40:55.909621 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h_6f893fd6-e5cc-42b9-ac46-77567818c0ef/manager/0.log" Oct 08 15:40:56 crc kubenswrapper[4789]: I1008 15:40:56.034533 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-766b688d6c-zfshx_d174fba3-84a1-4de8-8b32-4f97d1880a78/kube-rbac-proxy/0.log" Oct 08 15:40:56 crc kubenswrapper[4789]: I1008 15:40:56.182883 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5cf8d5fc8d-4524r_9eaa64d7-4d19-436b-a954-ffd75b6d6cfd/kube-rbac-proxy/0.log" Oct 08 15:40:56 crc kubenswrapper[4789]: I1008 15:40:56.391168 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-swzf7_cb5f8cb4-1588-4ae9-88b1-890831151eb4/registry-server/0.log" Oct 08 15:40:56 crc kubenswrapper[4789]: I1008 15:40:56.466628 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5cf8d5fc8d-4524r_9eaa64d7-4d19-436b-a954-ffd75b6d6cfd/operator/0.log" Oct 08 15:40:56 crc kubenswrapper[4789]: I1008 15:40:56.608389 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-6f9x9_5578f70a-ad1b-47f8-a6ce-5dc726788ded/kube-rbac-proxy/0.log" Oct 08 15:40:56 crc kubenswrapper[4789]: I1008 15:40:56.809680 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-6f9x9_5578f70a-ad1b-47f8-a6ce-5dc726788ded/manager/0.log" Oct 08 15:40:56 crc kubenswrapper[4789]: I1008 15:40:56.836802 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-4xgrb_67d8c63c-cc55-4be0-bea4-8371ac8e333b/kube-rbac-proxy/0.log" Oct 08 15:40:56 crc kubenswrapper[4789]: I1008 15:40:56.885442 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-4xgrb_67d8c63c-cc55-4be0-bea4-8371ac8e333b/manager/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.087320 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-nfllk_bc485c37-780a-4509-b3f3-aaf1714a9edb/kube-rbac-proxy/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.097060 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5_8bc03843-9303-451e-b933-c450aaad7401/operator/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.233479 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-nfllk_bc485c37-780a-4509-b3f3-aaf1714a9edb/manager/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.294300 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-775776c574-mgjfp_8efa82a4-efc9-4abe-b373-8828def5bb61/kube-rbac-proxy/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.516843 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-74665f6cdc-x52cc_08eb1945-dc5b-4f43-9fad-cd2b7e3102be/kube-rbac-proxy/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.569443 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-74665f6cdc-x52cc_08eb1945-dc5b-4f43-9fad-cd2b7e3102be/manager/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.692189 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-766b688d6c-zfshx_d174fba3-84a1-4de8-8b32-4f97d1880a78/manager/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.701128 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-775776c574-mgjfp_8efa82a4-efc9-4abe-b373-8828def5bb61/manager/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.718900 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c7dbfbcbd-7xrj5_8bdbf6d1-353c-42ba-80df-e60d6b424f55/kube-rbac-proxy/0.log" Oct 08 15:40:57 crc kubenswrapper[4789]: I1008 15:40:57.836254 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c7dbfbcbd-7xrj5_8bdbf6d1-353c-42ba-80df-e60d6b424f55/manager/0.log" Oct 08 15:40:59 crc kubenswrapper[4789]: I1008 15:40:59.730078 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:40:59 crc kubenswrapper[4789]: E1008 15:40:59.730899 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.165788 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b8tgw"] Oct 08 15:41:02 crc kubenswrapper[4789]: E1008 15:41:02.166887 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="334a1212-ea46-4ef4-8770-93ce8c09dfce" containerName="container-00" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.166908 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="334a1212-ea46-4ef4-8770-93ce8c09dfce" containerName="container-00" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.167210 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="334a1212-ea46-4ef4-8770-93ce8c09dfce" containerName="container-00" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.169054 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.177606 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8tgw"] Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.264569 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5pjt\" (UniqueName: \"kubernetes.io/projected/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-kube-api-access-h5pjt\") pod \"redhat-marketplace-b8tgw\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.266072 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-catalog-content\") pod \"redhat-marketplace-b8tgw\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.266278 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-utilities\") pod \"redhat-marketplace-b8tgw\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.367944 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5pjt\" (UniqueName: \"kubernetes.io/projected/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-kube-api-access-h5pjt\") pod \"redhat-marketplace-b8tgw\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.368046 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-catalog-content\") pod \"redhat-marketplace-b8tgw\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.368084 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-utilities\") pod \"redhat-marketplace-b8tgw\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.368672 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-utilities\") pod \"redhat-marketplace-b8tgw\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.368846 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-catalog-content\") pod \"redhat-marketplace-b8tgw\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.389819 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5pjt\" (UniqueName: \"kubernetes.io/projected/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-kube-api-access-h5pjt\") pod \"redhat-marketplace-b8tgw\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:02 crc kubenswrapper[4789]: I1008 15:41:02.507691 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:03 crc kubenswrapper[4789]: I1008 15:41:03.010950 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8tgw"] Oct 08 15:41:03 crc kubenswrapper[4789]: I1008 15:41:03.262618 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8tgw" event={"ID":"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7","Type":"ContainerStarted","Data":"c7da6f625bd653c105ec21701893fd6aee10a3aa7dde34c91e80137ab6ff498c"} Oct 08 15:41:04 crc kubenswrapper[4789]: I1008 15:41:04.280654 4789 generic.go:334] "Generic (PLEG): container finished" podID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerID="5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e" exitCode=0 Oct 08 15:41:04 crc kubenswrapper[4789]: I1008 15:41:04.280743 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8tgw" event={"ID":"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7","Type":"ContainerDied","Data":"5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e"} Oct 08 15:41:05 crc kubenswrapper[4789]: I1008 15:41:05.292359 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8tgw" event={"ID":"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7","Type":"ContainerStarted","Data":"7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2"} Oct 08 15:41:06 crc kubenswrapper[4789]: I1008 15:41:06.308078 4789 generic.go:334] "Generic (PLEG): container finished" podID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerID="7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2" exitCode=0 Oct 08 15:41:06 crc kubenswrapper[4789]: I1008 15:41:06.308132 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8tgw" event={"ID":"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7","Type":"ContainerDied","Data":"7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2"} Oct 08 15:41:07 crc kubenswrapper[4789]: I1008 15:41:07.318612 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8tgw" event={"ID":"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7","Type":"ContainerStarted","Data":"274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97"} Oct 08 15:41:07 crc kubenswrapper[4789]: I1008 15:41:07.334927 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b8tgw" podStartSLOduration=2.818453712 podStartE2EDuration="5.334909227s" podCreationTimestamp="2025-10-08 15:41:02 +0000 UTC" firstStartedPulling="2025-10-08 15:41:04.284366661 +0000 UTC m=+6004.191114153" lastFinishedPulling="2025-10-08 15:41:06.800822176 +0000 UTC m=+6006.707569668" observedRunningTime="2025-10-08 15:41:07.332804689 +0000 UTC m=+6007.239552181" watchObservedRunningTime="2025-10-08 15:41:07.334909227 +0000 UTC m=+6007.241656719" Oct 08 15:41:10 crc kubenswrapper[4789]: I1008 15:41:10.739319 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:41:10 crc kubenswrapper[4789]: E1008 15:41:10.742265 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:41:12 crc kubenswrapper[4789]: I1008 15:41:12.508420 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:12 crc kubenswrapper[4789]: I1008 15:41:12.508783 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:12 crc kubenswrapper[4789]: I1008 15:41:12.556632 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:13 crc kubenswrapper[4789]: I1008 15:41:13.199652 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-c8rg8_cbd6f891-627b-4f0b-ac4d-40e30d01fb89/control-plane-machine-set-operator/0.log" Oct 08 15:41:13 crc kubenswrapper[4789]: I1008 15:41:13.412319 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wmqxc_bc2cc553-6ac7-44a6-948c-c1363836923d/kube-rbac-proxy/0.log" Oct 08 15:41:13 crc kubenswrapper[4789]: I1008 15:41:13.429686 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wmqxc_bc2cc553-6ac7-44a6-948c-c1363836923d/machine-api-operator/0.log" Oct 08 15:41:13 crc kubenswrapper[4789]: I1008 15:41:13.443271 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:13 crc kubenswrapper[4789]: I1008 15:41:13.496155 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8tgw"] Oct 08 15:41:15 crc kubenswrapper[4789]: I1008 15:41:15.407116 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b8tgw" podUID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerName="registry-server" containerID="cri-o://274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97" gracePeriod=2 Oct 08 15:41:15 crc kubenswrapper[4789]: I1008 15:41:15.937532 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.055132 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-catalog-content\") pod \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.055339 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5pjt\" (UniqueName: \"kubernetes.io/projected/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-kube-api-access-h5pjt\") pod \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.055457 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-utilities\") pod \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\" (UID: \"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7\") " Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.057080 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-utilities" (OuterVolumeSpecName: "utilities") pod "6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" (UID: "6e0061aa-27e8-41ec-b79e-cf701ce5a1a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.074230 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-kube-api-access-h5pjt" (OuterVolumeSpecName: "kube-api-access-h5pjt") pod "6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" (UID: "6e0061aa-27e8-41ec-b79e-cf701ce5a1a7"). InnerVolumeSpecName "kube-api-access-h5pjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.117218 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" (UID: "6e0061aa-27e8-41ec-b79e-cf701ce5a1a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.159421 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.159469 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5pjt\" (UniqueName: \"kubernetes.io/projected/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-kube-api-access-h5pjt\") on node \"crc\" DevicePath \"\"" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.159482 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.417088 4789 generic.go:334] "Generic (PLEG): container finished" podID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerID="274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97" exitCode=0 Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.417150 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8tgw" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.417167 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8tgw" event={"ID":"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7","Type":"ContainerDied","Data":"274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97"} Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.418230 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8tgw" event={"ID":"6e0061aa-27e8-41ec-b79e-cf701ce5a1a7","Type":"ContainerDied","Data":"c7da6f625bd653c105ec21701893fd6aee10a3aa7dde34c91e80137ab6ff498c"} Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.418254 4789 scope.go:117] "RemoveContainer" containerID="274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.440790 4789 scope.go:117] "RemoveContainer" containerID="7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.465299 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8tgw"] Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.474573 4789 scope.go:117] "RemoveContainer" containerID="5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.483671 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8tgw"] Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.510193 4789 scope.go:117] "RemoveContainer" containerID="274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97" Oct 08 15:41:16 crc kubenswrapper[4789]: E1008 15:41:16.510712 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97\": container with ID starting with 274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97 not found: ID does not exist" containerID="274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.510753 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97"} err="failed to get container status \"274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97\": rpc error: code = NotFound desc = could not find container \"274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97\": container with ID starting with 274984607c8b580edb6fe1af9893cb2c12fa2da458ef9a91dd87037eb136eb97 not found: ID does not exist" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.510782 4789 scope.go:117] "RemoveContainer" containerID="7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2" Oct 08 15:41:16 crc kubenswrapper[4789]: E1008 15:41:16.511231 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2\": container with ID starting with 7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2 not found: ID does not exist" containerID="7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.511270 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2"} err="failed to get container status \"7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2\": rpc error: code = NotFound desc = could not find container \"7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2\": container with ID starting with 7f6aa9f5fd219c4992126e9889e4b8dd7afd4687ef1dee6038586fc18a583cf2 not found: ID does not exist" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.511296 4789 scope.go:117] "RemoveContainer" containerID="5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e" Oct 08 15:41:16 crc kubenswrapper[4789]: E1008 15:41:16.511640 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e\": container with ID starting with 5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e not found: ID does not exist" containerID="5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.511673 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e"} err="failed to get container status \"5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e\": rpc error: code = NotFound desc = could not find container \"5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e\": container with ID starting with 5bf2a9cefdf79a046e46422585a88f494add5a19d67efe7232d6d223645dd95e not found: ID does not exist" Oct 08 15:41:16 crc kubenswrapper[4789]: I1008 15:41:16.740901 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" path="/var/lib/kubelet/pods/6e0061aa-27e8-41ec-b79e-cf701ce5a1a7/volumes" Oct 08 15:41:24 crc kubenswrapper[4789]: I1008 15:41:24.730065 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:41:24 crc kubenswrapper[4789]: E1008 15:41:24.730893 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:41:25 crc kubenswrapper[4789]: I1008 15:41:25.094305 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-qtnnc_a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d/cert-manager-controller/0.log" Oct 08 15:41:25 crc kubenswrapper[4789]: I1008 15:41:25.217191 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-5vfnw_e05b78ba-5b41-4c35-b613-87ff25afd63b/cert-manager-cainjector/0.log" Oct 08 15:41:25 crc kubenswrapper[4789]: I1008 15:41:25.275147 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dkqb4_cf871183-8c76-4ab3-af84-a483ad387b7b/cert-manager-webhook/0.log" Oct 08 15:41:35 crc kubenswrapper[4789]: I1008 15:41:35.730256 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:41:35 crc kubenswrapper[4789]: E1008 15:41:35.731146 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:41:35 crc kubenswrapper[4789]: I1008 15:41:35.987483 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-4cxwh_b5615e4f-312c-49dc-9689-3662e0ad7ac8/nmstate-console-plugin/0.log" Oct 08 15:41:36 crc kubenswrapper[4789]: I1008 15:41:36.182324 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2qnx8_764cebf9-8453-4c3b-afe4-c6f56b39d87f/nmstate-handler/0.log" Oct 08 15:41:36 crc kubenswrapper[4789]: I1008 15:41:36.288089 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-l7cdj_e278e129-8d89-4095-a7f0-88a44cb19c97/nmstate-metrics/0.log" Oct 08 15:41:36 crc kubenswrapper[4789]: I1008 15:41:36.311539 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-l7cdj_e278e129-8d89-4095-a7f0-88a44cb19c97/kube-rbac-proxy/0.log" Oct 08 15:41:36 crc kubenswrapper[4789]: I1008 15:41:36.435153 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-cbqs2_333279f2-ccf6-44df-80ad-9097f320a3ac/nmstate-operator/0.log" Oct 08 15:41:36 crc kubenswrapper[4789]: I1008 15:41:36.499974 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-8ndgl_066fb44b-7e80-4cea-b5c9-ab9e208a9d52/nmstate-webhook/0.log" Oct 08 15:41:50 crc kubenswrapper[4789]: I1008 15:41:50.435571 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-ggzfl_9f942609-b5dc-4433-8767-a9bd36b185f9/kube-rbac-proxy/0.log" Oct 08 15:41:50 crc kubenswrapper[4789]: I1008 15:41:50.723028 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-ggzfl_9f942609-b5dc-4433-8767-a9bd36b185f9/controller/0.log" Oct 08 15:41:50 crc kubenswrapper[4789]: I1008 15:41:50.742226 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:41:50 crc kubenswrapper[4789]: E1008 15:41:50.742484 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:41:50 crc kubenswrapper[4789]: I1008 15:41:50.768915 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-frr-files/0.log" Oct 08 15:41:50 crc kubenswrapper[4789]: I1008 15:41:50.932046 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-frr-files/0.log" Oct 08 15:41:50 crc kubenswrapper[4789]: I1008 15:41:50.949001 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-reloader/0.log" Oct 08 15:41:50 crc kubenswrapper[4789]: I1008 15:41:50.997804 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-reloader/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.002634 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-metrics/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.172949 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-frr-files/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.193425 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-metrics/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.237263 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-reloader/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.279113 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-metrics/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.414287 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-frr-files/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.436078 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-metrics/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.485386 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-reloader/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.493342 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/controller/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.671269 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/frr-metrics/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.672340 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/kube-rbac-proxy/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.730147 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/kube-rbac-proxy-frr/0.log" Oct 08 15:41:51 crc kubenswrapper[4789]: I1008 15:41:51.942533 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/reloader/0.log" Oct 08 15:41:52 crc kubenswrapper[4789]: I1008 15:41:52.022317 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-qx7cs_988fe042-8dd6-4398-af3d-ae201442f223/frr-k8s-webhook-server/0.log" Oct 08 15:41:52 crc kubenswrapper[4789]: I1008 15:41:52.277227 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-69d56bf45b-9wq7v_222ee176-36f5-4f7e-b6cf-8c526a0008d6/manager/0.log" Oct 08 15:41:52 crc kubenswrapper[4789]: I1008 15:41:52.428447 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-655bc78868-jzwhv_067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd/webhook-server/0.log" Oct 08 15:41:52 crc kubenswrapper[4789]: I1008 15:41:52.611463 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-646mh_b76ec5f4-3801-4d1a-9b64-5b7beebc4100/kube-rbac-proxy/0.log" Oct 08 15:41:53 crc kubenswrapper[4789]: I1008 15:41:53.352642 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/frr/0.log" Oct 08 15:41:53 crc kubenswrapper[4789]: I1008 15:41:53.362616 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-646mh_b76ec5f4-3801-4d1a-9b64-5b7beebc4100/speaker/0.log" Oct 08 15:42:04 crc kubenswrapper[4789]: I1008 15:42:04.730470 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:42:05 crc kubenswrapper[4789]: I1008 15:42:05.601889 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/util/0.log" Oct 08 15:42:05 crc kubenswrapper[4789]: I1008 15:42:05.874585 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"bd9daa530e0df0e138a935950ee0c210dd40007a2821f368084c00d54709c8aa"} Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.007102 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/pull/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.013616 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/pull/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.032061 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/util/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.262455 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/util/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.262482 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/pull/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.263109 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/extract/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.453868 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/util/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.627411 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/util/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.662925 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/pull/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.676488 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/pull/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.850090 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/util/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.868218 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/pull/0.log" Oct 08 15:42:06 crc kubenswrapper[4789]: I1008 15:42:06.889477 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/extract/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.090050 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-utilities/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.254033 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-utilities/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.261958 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-content/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.284665 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-content/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.436933 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-utilities/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.437797 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-content/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.706715 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/registry-server/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.715318 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-utilities/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.861931 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-utilities/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.906460 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-content/0.log" Oct 08 15:42:07 crc kubenswrapper[4789]: I1008 15:42:07.942655 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-content/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.079808 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-content/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.107566 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-utilities/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.402190 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/registry-server/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.413360 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/util/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.558300 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/pull/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.568734 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/util/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.606879 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/pull/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.774307 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/util/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.810035 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/extract/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.814457 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/pull/0.log" Oct 08 15:42:08 crc kubenswrapper[4789]: I1008 15:42:08.978024 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4lxzv_c59c8d39-3840-48cf-b510-4e07d9293fc9/marketplace-operator/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.038283 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-utilities/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.259940 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-content/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.269823 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-content/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.269868 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-utilities/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.444235 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-utilities/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.487166 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-utilities/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.489701 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-content/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.727839 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/registry-server/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.733930 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-content/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.734032 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-utilities/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.769892 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-content/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.914169 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-content/0.log" Oct 08 15:42:09 crc kubenswrapper[4789]: I1008 15:42:09.920957 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-utilities/0.log" Oct 08 15:42:10 crc kubenswrapper[4789]: I1008 15:42:10.851934 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/registry-server/0.log" Oct 08 15:42:21 crc kubenswrapper[4789]: I1008 15:42:21.381921 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-lwgzt_5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec/prometheus-operator/0.log" Oct 08 15:42:21 crc kubenswrapper[4789]: I1008 15:42:21.542640 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l_9451e6d9-fa5b-4209-a332-a65412994d55/prometheus-operator-admission-webhook/0.log" Oct 08 15:42:21 crc kubenswrapper[4789]: I1008 15:42:21.615678 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh_7141b3e4-ea4a-4476-84d2-57670fe5e932/prometheus-operator-admission-webhook/0.log" Oct 08 15:42:21 crc kubenswrapper[4789]: I1008 15:42:21.757379 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-tclgl_643a32e3-75dd-44f5-bbac-e2d917641a29/operator/0.log" Oct 08 15:42:21 crc kubenswrapper[4789]: I1008 15:42:21.818598 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-rbwl7_1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5/perses-operator/0.log" Oct 08 15:44:26 crc kubenswrapper[4789]: I1008 15:44:26.433155 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:44:26 crc kubenswrapper[4789]: I1008 15:44:26.433850 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:44:26 crc kubenswrapper[4789]: I1008 15:44:26.643553 4789 scope.go:117] "RemoveContainer" containerID="8625cabf3cfb56081793dbdc93ef4a921f6f2f59d7a3b54a98b9dc77dca933b3" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.361778 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dzkgj"] Oct 08 15:44:27 crc kubenswrapper[4789]: E1008 15:44:27.362616 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerName="extract-content" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.362636 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerName="extract-content" Oct 08 15:44:27 crc kubenswrapper[4789]: E1008 15:44:27.362648 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerName="registry-server" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.362655 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerName="registry-server" Oct 08 15:44:27 crc kubenswrapper[4789]: E1008 15:44:27.362691 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerName="extract-utilities" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.362698 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerName="extract-utilities" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.362902 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0061aa-27e8-41ec-b79e-cf701ce5a1a7" containerName="registry-server" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.364833 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.397134 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dzkgj"] Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.476964 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-catalog-content\") pod \"community-operators-dzkgj\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.477047 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-utilities\") pod \"community-operators-dzkgj\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.477207 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rl2m\" (UniqueName: \"kubernetes.io/projected/48492e54-a9ca-4167-b7dc-fe68c62c1954-kube-api-access-4rl2m\") pod \"community-operators-dzkgj\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.579004 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-catalog-content\") pod \"community-operators-dzkgj\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.579333 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-utilities\") pod \"community-operators-dzkgj\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.579413 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rl2m\" (UniqueName: \"kubernetes.io/projected/48492e54-a9ca-4167-b7dc-fe68c62c1954-kube-api-access-4rl2m\") pod \"community-operators-dzkgj\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.579587 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-catalog-content\") pod \"community-operators-dzkgj\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.579887 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-utilities\") pod \"community-operators-dzkgj\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.609913 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rl2m\" (UniqueName: \"kubernetes.io/projected/48492e54-a9ca-4167-b7dc-fe68c62c1954-kube-api-access-4rl2m\") pod \"community-operators-dzkgj\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:27 crc kubenswrapper[4789]: I1008 15:44:27.685212 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:28 crc kubenswrapper[4789]: I1008 15:44:28.261595 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dzkgj"] Oct 08 15:44:28 crc kubenswrapper[4789]: I1008 15:44:28.358801 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzkgj" event={"ID":"48492e54-a9ca-4167-b7dc-fe68c62c1954","Type":"ContainerStarted","Data":"5f4655f3ae3056a97c76f3828c72b078dd36ed8de25c29fc865e4d3ca5199f8f"} Oct 08 15:44:29 crc kubenswrapper[4789]: I1008 15:44:29.369187 4789 generic.go:334] "Generic (PLEG): container finished" podID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerID="bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4" exitCode=0 Oct 08 15:44:29 crc kubenswrapper[4789]: I1008 15:44:29.369309 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzkgj" event={"ID":"48492e54-a9ca-4167-b7dc-fe68c62c1954","Type":"ContainerDied","Data":"bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4"} Oct 08 15:44:29 crc kubenswrapper[4789]: I1008 15:44:29.371776 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:44:31 crc kubenswrapper[4789]: I1008 15:44:31.388706 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzkgj" event={"ID":"48492e54-a9ca-4167-b7dc-fe68c62c1954","Type":"ContainerStarted","Data":"594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939"} Oct 08 15:44:32 crc kubenswrapper[4789]: I1008 15:44:32.402545 4789 generic.go:334] "Generic (PLEG): container finished" podID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerID="594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939" exitCode=0 Oct 08 15:44:32 crc kubenswrapper[4789]: I1008 15:44:32.402603 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzkgj" event={"ID":"48492e54-a9ca-4167-b7dc-fe68c62c1954","Type":"ContainerDied","Data":"594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939"} Oct 08 15:44:33 crc kubenswrapper[4789]: I1008 15:44:33.413897 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzkgj" event={"ID":"48492e54-a9ca-4167-b7dc-fe68c62c1954","Type":"ContainerStarted","Data":"bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933"} Oct 08 15:44:33 crc kubenswrapper[4789]: I1008 15:44:33.433922 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dzkgj" podStartSLOduration=2.964149001 podStartE2EDuration="6.433906957s" podCreationTimestamp="2025-10-08 15:44:27 +0000 UTC" firstStartedPulling="2025-10-08 15:44:29.371471684 +0000 UTC m=+6209.278219176" lastFinishedPulling="2025-10-08 15:44:32.84122964 +0000 UTC m=+6212.747977132" observedRunningTime="2025-10-08 15:44:33.428692345 +0000 UTC m=+6213.335439827" watchObservedRunningTime="2025-10-08 15:44:33.433906957 +0000 UTC m=+6213.340654449" Oct 08 15:44:37 crc kubenswrapper[4789]: I1008 15:44:37.685365 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:37 crc kubenswrapper[4789]: I1008 15:44:37.686896 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:37 crc kubenswrapper[4789]: I1008 15:44:37.737114 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:38 crc kubenswrapper[4789]: I1008 15:44:38.503349 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:38 crc kubenswrapper[4789]: I1008 15:44:38.551201 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dzkgj"] Oct 08 15:44:40 crc kubenswrapper[4789]: I1008 15:44:40.470872 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dzkgj" podUID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerName="registry-server" containerID="cri-o://bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933" gracePeriod=2 Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.167626 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.177285 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-utilities\") pod \"48492e54-a9ca-4167-b7dc-fe68c62c1954\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.177747 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-catalog-content\") pod \"48492e54-a9ca-4167-b7dc-fe68c62c1954\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.177780 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rl2m\" (UniqueName: \"kubernetes.io/projected/48492e54-a9ca-4167-b7dc-fe68c62c1954-kube-api-access-4rl2m\") pod \"48492e54-a9ca-4167-b7dc-fe68c62c1954\" (UID: \"48492e54-a9ca-4167-b7dc-fe68c62c1954\") " Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.179135 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-utilities" (OuterVolumeSpecName: "utilities") pod "48492e54-a9ca-4167-b7dc-fe68c62c1954" (UID: "48492e54-a9ca-4167-b7dc-fe68c62c1954"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.184287 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48492e54-a9ca-4167-b7dc-fe68c62c1954-kube-api-access-4rl2m" (OuterVolumeSpecName: "kube-api-access-4rl2m") pod "48492e54-a9ca-4167-b7dc-fe68c62c1954" (UID: "48492e54-a9ca-4167-b7dc-fe68c62c1954"). InnerVolumeSpecName "kube-api-access-4rl2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.280356 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.280390 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rl2m\" (UniqueName: \"kubernetes.io/projected/48492e54-a9ca-4167-b7dc-fe68c62c1954-kube-api-access-4rl2m\") on node \"crc\" DevicePath \"\"" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.481611 4789 generic.go:334] "Generic (PLEG): container finished" podID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerID="bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933" exitCode=0 Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.481665 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzkgj" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.481688 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzkgj" event={"ID":"48492e54-a9ca-4167-b7dc-fe68c62c1954","Type":"ContainerDied","Data":"bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933"} Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.482881 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzkgj" event={"ID":"48492e54-a9ca-4167-b7dc-fe68c62c1954","Type":"ContainerDied","Data":"5f4655f3ae3056a97c76f3828c72b078dd36ed8de25c29fc865e4d3ca5199f8f"} Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.482928 4789 scope.go:117] "RemoveContainer" containerID="bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.501541 4789 scope.go:117] "RemoveContainer" containerID="594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.528051 4789 scope.go:117] "RemoveContainer" containerID="bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.576860 4789 scope.go:117] "RemoveContainer" containerID="bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933" Oct 08 15:44:41 crc kubenswrapper[4789]: E1008 15:44:41.577405 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933\": container with ID starting with bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933 not found: ID does not exist" containerID="bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.577444 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933"} err="failed to get container status \"bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933\": rpc error: code = NotFound desc = could not find container \"bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933\": container with ID starting with bdc1e32877f698978b9eb2396547b02a083c4d58fbf4b7a9d4efe5b5ae0ad933 not found: ID does not exist" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.577471 4789 scope.go:117] "RemoveContainer" containerID="594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939" Oct 08 15:44:41 crc kubenswrapper[4789]: E1008 15:44:41.577808 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939\": container with ID starting with 594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939 not found: ID does not exist" containerID="594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.577852 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939"} err="failed to get container status \"594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939\": rpc error: code = NotFound desc = could not find container \"594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939\": container with ID starting with 594504a1b9fd4f0769533e9121ffef938c32b96cc90ac367faaad92f55e2a939 not found: ID does not exist" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.577884 4789 scope.go:117] "RemoveContainer" containerID="bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4" Oct 08 15:44:41 crc kubenswrapper[4789]: E1008 15:44:41.578323 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4\": container with ID starting with bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4 not found: ID does not exist" containerID="bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4" Oct 08 15:44:41 crc kubenswrapper[4789]: I1008 15:44:41.578373 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4"} err="failed to get container status \"bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4\": rpc error: code = NotFound desc = could not find container \"bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4\": container with ID starting with bcba124cc3e8f31f526cb0d89bc134056f6f18f96a880f0d3fb2146e37a89cd4 not found: ID does not exist" Oct 08 15:44:42 crc kubenswrapper[4789]: I1008 15:44:42.097973 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48492e54-a9ca-4167-b7dc-fe68c62c1954" (UID: "48492e54-a9ca-4167-b7dc-fe68c62c1954"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:44:42 crc kubenswrapper[4789]: I1008 15:44:42.196967 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48492e54-a9ca-4167-b7dc-fe68c62c1954-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:44:42 crc kubenswrapper[4789]: I1008 15:44:42.435255 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dzkgj"] Oct 08 15:44:42 crc kubenswrapper[4789]: I1008 15:44:42.443747 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dzkgj"] Oct 08 15:44:42 crc kubenswrapper[4789]: I1008 15:44:42.496295 4789 generic.go:334] "Generic (PLEG): container finished" podID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" containerID="97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8" exitCode=0 Oct 08 15:44:42 crc kubenswrapper[4789]: I1008 15:44:42.496341 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" event={"ID":"12cb06f8-6eb0-4468-ac4b-257d2e36b765","Type":"ContainerDied","Data":"97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8"} Oct 08 15:44:42 crc kubenswrapper[4789]: I1008 15:44:42.497012 4789 scope.go:117] "RemoveContainer" containerID="97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8" Oct 08 15:44:42 crc kubenswrapper[4789]: I1008 15:44:42.740804 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48492e54-a9ca-4167-b7dc-fe68c62c1954" path="/var/lib/kubelet/pods/48492e54-a9ca-4167-b7dc-fe68c62c1954/volumes" Oct 08 15:44:43 crc kubenswrapper[4789]: I1008 15:44:43.202579 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gqlwf_must-gather-hq4p2_12cb06f8-6eb0-4468-ac4b-257d2e36b765/gather/0.log" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.046051 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gqlwf/must-gather-hq4p2"] Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.047307 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" podUID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" containerName="copy" containerID="cri-o://d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772" gracePeriod=2 Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.058974 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gqlwf/must-gather-hq4p2"] Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.522844 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gqlwf_must-gather-hq4p2_12cb06f8-6eb0-4468-ac4b-257d2e36b765/copy/0.log" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.523517 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.597607 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gqlwf_must-gather-hq4p2_12cb06f8-6eb0-4468-ac4b-257d2e36b765/copy/0.log" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.598480 4789 generic.go:334] "Generic (PLEG): container finished" podID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" containerID="d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772" exitCode=143 Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.598550 4789 scope.go:117] "RemoveContainer" containerID="d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.598575 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqlwf/must-gather-hq4p2" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.623345 4789 scope.go:117] "RemoveContainer" containerID="97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.641701 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmhrl\" (UniqueName: \"kubernetes.io/projected/12cb06f8-6eb0-4468-ac4b-257d2e36b765-kube-api-access-kmhrl\") pod \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\" (UID: \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\") " Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.641865 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12cb06f8-6eb0-4468-ac4b-257d2e36b765-must-gather-output\") pod \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\" (UID: \"12cb06f8-6eb0-4468-ac4b-257d2e36b765\") " Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.656582 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12cb06f8-6eb0-4468-ac4b-257d2e36b765-kube-api-access-kmhrl" (OuterVolumeSpecName: "kube-api-access-kmhrl") pod "12cb06f8-6eb0-4468-ac4b-257d2e36b765" (UID: "12cb06f8-6eb0-4468-ac4b-257d2e36b765"). InnerVolumeSpecName "kube-api-access-kmhrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.680473 4789 scope.go:117] "RemoveContainer" containerID="d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772" Oct 08 15:44:52 crc kubenswrapper[4789]: E1008 15:44:52.681941 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772\": container with ID starting with d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772 not found: ID does not exist" containerID="d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.681972 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772"} err="failed to get container status \"d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772\": rpc error: code = NotFound desc = could not find container \"d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772\": container with ID starting with d76d9f7bb207dfdd487b4f9a617da7a0c1eb50dd37692116490740b80f1fc772 not found: ID does not exist" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.682009 4789 scope.go:117] "RemoveContainer" containerID="97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8" Oct 08 15:44:52 crc kubenswrapper[4789]: E1008 15:44:52.682321 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8\": container with ID starting with 97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8 not found: ID does not exist" containerID="97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.682345 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8"} err="failed to get container status \"97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8\": rpc error: code = NotFound desc = could not find container \"97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8\": container with ID starting with 97033c14500209441b41a40353421a692b00098f5c1ba41f68e94834d9969ed8 not found: ID does not exist" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.744263 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmhrl\" (UniqueName: \"kubernetes.io/projected/12cb06f8-6eb0-4468-ac4b-257d2e36b765-kube-api-access-kmhrl\") on node \"crc\" DevicePath \"\"" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.876686 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12cb06f8-6eb0-4468-ac4b-257d2e36b765-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "12cb06f8-6eb0-4468-ac4b-257d2e36b765" (UID: "12cb06f8-6eb0-4468-ac4b-257d2e36b765"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:44:52 crc kubenswrapper[4789]: I1008 15:44:52.950405 4789 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/12cb06f8-6eb0-4468-ac4b-257d2e36b765-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 15:44:54 crc kubenswrapper[4789]: I1008 15:44:54.741822 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" path="/var/lib/kubelet/pods/12cb06f8-6eb0-4468-ac4b-257d2e36b765/volumes" Oct 08 15:44:56 crc kubenswrapper[4789]: I1008 15:44:56.432909 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:44:56 crc kubenswrapper[4789]: I1008 15:44:56.433747 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.162868 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d"] Oct 08 15:45:00 crc kubenswrapper[4789]: E1008 15:45:00.164782 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" containerName="gather" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.164874 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" containerName="gather" Oct 08 15:45:00 crc kubenswrapper[4789]: E1008 15:45:00.164958 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerName="registry-server" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.165053 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerName="registry-server" Oct 08 15:45:00 crc kubenswrapper[4789]: E1008 15:45:00.165128 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerName="extract-content" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.165192 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerName="extract-content" Oct 08 15:45:00 crc kubenswrapper[4789]: E1008 15:45:00.165264 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerName="extract-utilities" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.165342 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerName="extract-utilities" Oct 08 15:45:00 crc kubenswrapper[4789]: E1008 15:45:00.165426 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" containerName="copy" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.165487 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" containerName="copy" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.165752 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" containerName="copy" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.165843 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="12cb06f8-6eb0-4468-ac4b-257d2e36b765" containerName="gather" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.165924 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="48492e54-a9ca-4167-b7dc-fe68c62c1954" containerName="registry-server" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.166728 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.170520 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.173519 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.177565 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d"] Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.206754 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e358a26b-0e6c-4f91-971b-a684fcd7c502-config-volume\") pod \"collect-profiles-29332305-x2w8d\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.206813 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8c4q\" (UniqueName: \"kubernetes.io/projected/e358a26b-0e6c-4f91-971b-a684fcd7c502-kube-api-access-t8c4q\") pod \"collect-profiles-29332305-x2w8d\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.206839 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e358a26b-0e6c-4f91-971b-a684fcd7c502-secret-volume\") pod \"collect-profiles-29332305-x2w8d\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.308835 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e358a26b-0e6c-4f91-971b-a684fcd7c502-config-volume\") pod \"collect-profiles-29332305-x2w8d\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.308903 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8c4q\" (UniqueName: \"kubernetes.io/projected/e358a26b-0e6c-4f91-971b-a684fcd7c502-kube-api-access-t8c4q\") pod \"collect-profiles-29332305-x2w8d\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.308926 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e358a26b-0e6c-4f91-971b-a684fcd7c502-secret-volume\") pod \"collect-profiles-29332305-x2w8d\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.309827 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e358a26b-0e6c-4f91-971b-a684fcd7c502-config-volume\") pod \"collect-profiles-29332305-x2w8d\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.321727 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e358a26b-0e6c-4f91-971b-a684fcd7c502-secret-volume\") pod \"collect-profiles-29332305-x2w8d\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.330367 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8c4q\" (UniqueName: \"kubernetes.io/projected/e358a26b-0e6c-4f91-971b-a684fcd7c502-kube-api-access-t8c4q\") pod \"collect-profiles-29332305-x2w8d\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.487172 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:00 crc kubenswrapper[4789]: I1008 15:45:00.966494 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d"] Oct 08 15:45:01 crc kubenswrapper[4789]: I1008 15:45:01.696593 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" event={"ID":"e358a26b-0e6c-4f91-971b-a684fcd7c502","Type":"ContainerStarted","Data":"5a3e03b28f83451afa85a35e19b098f7f930addf1162135f4d7f5cfb1ff0a627"} Oct 08 15:45:01 crc kubenswrapper[4789]: I1008 15:45:01.697826 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" event={"ID":"e358a26b-0e6c-4f91-971b-a684fcd7c502","Type":"ContainerStarted","Data":"e37ecb5b0076b9f7827b6c62f36ae91734b3a4fa4a74342ad3b81c85375773ad"} Oct 08 15:45:01 crc kubenswrapper[4789]: I1008 15:45:01.716389 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" podStartSLOduration=1.716369425 podStartE2EDuration="1.716369425s" podCreationTimestamp="2025-10-08 15:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:45:01.708604183 +0000 UTC m=+6241.615351675" watchObservedRunningTime="2025-10-08 15:45:01.716369425 +0000 UTC m=+6241.623116917" Oct 08 15:45:02 crc kubenswrapper[4789]: I1008 15:45:02.708689 4789 generic.go:334] "Generic (PLEG): container finished" podID="e358a26b-0e6c-4f91-971b-a684fcd7c502" containerID="5a3e03b28f83451afa85a35e19b098f7f930addf1162135f4d7f5cfb1ff0a627" exitCode=0 Oct 08 15:45:02 crc kubenswrapper[4789]: I1008 15:45:02.708783 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" event={"ID":"e358a26b-0e6c-4f91-971b-a684fcd7c502","Type":"ContainerDied","Data":"5a3e03b28f83451afa85a35e19b098f7f930addf1162135f4d7f5cfb1ff0a627"} Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.105167 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.192029 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e358a26b-0e6c-4f91-971b-a684fcd7c502-config-volume\") pod \"e358a26b-0e6c-4f91-971b-a684fcd7c502\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.192235 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8c4q\" (UniqueName: \"kubernetes.io/projected/e358a26b-0e6c-4f91-971b-a684fcd7c502-kube-api-access-t8c4q\") pod \"e358a26b-0e6c-4f91-971b-a684fcd7c502\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.192297 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e358a26b-0e6c-4f91-971b-a684fcd7c502-secret-volume\") pod \"e358a26b-0e6c-4f91-971b-a684fcd7c502\" (UID: \"e358a26b-0e6c-4f91-971b-a684fcd7c502\") " Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.192863 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e358a26b-0e6c-4f91-971b-a684fcd7c502-config-volume" (OuterVolumeSpecName: "config-volume") pod "e358a26b-0e6c-4f91-971b-a684fcd7c502" (UID: "e358a26b-0e6c-4f91-971b-a684fcd7c502"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.199194 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e358a26b-0e6c-4f91-971b-a684fcd7c502-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e358a26b-0e6c-4f91-971b-a684fcd7c502" (UID: "e358a26b-0e6c-4f91-971b-a684fcd7c502"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.202278 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e358a26b-0e6c-4f91-971b-a684fcd7c502-kube-api-access-t8c4q" (OuterVolumeSpecName: "kube-api-access-t8c4q") pod "e358a26b-0e6c-4f91-971b-a684fcd7c502" (UID: "e358a26b-0e6c-4f91-971b-a684fcd7c502"). InnerVolumeSpecName "kube-api-access-t8c4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.294567 4789 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e358a26b-0e6c-4f91-971b-a684fcd7c502-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.294633 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8c4q\" (UniqueName: \"kubernetes.io/projected/e358a26b-0e6c-4f91-971b-a684fcd7c502-kube-api-access-t8c4q\") on node \"crc\" DevicePath \"\"" Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.294646 4789 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e358a26b-0e6c-4f91-971b-a684fcd7c502-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.727621 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" event={"ID":"e358a26b-0e6c-4f91-971b-a684fcd7c502","Type":"ContainerDied","Data":"e37ecb5b0076b9f7827b6c62f36ae91734b3a4fa4a74342ad3b81c85375773ad"} Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.727651 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e37ecb5b0076b9f7827b6c62f36ae91734b3a4fa4a74342ad3b81c85375773ad" Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.727653 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332305-x2w8d" Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.796843 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2"] Oct 08 15:45:04 crc kubenswrapper[4789]: I1008 15:45:04.812013 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332260-rcsz2"] Oct 08 15:45:06 crc kubenswrapper[4789]: I1008 15:45:06.742446 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03d9e7f6-2d4e-4512-91ab-0430a401721d" path="/var/lib/kubelet/pods/03d9e7f6-2d4e-4512-91ab-0430a401721d/volumes" Oct 08 15:45:08 crc kubenswrapper[4789]: I1008 15:45:08.784668 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pj669"] Oct 08 15:45:08 crc kubenswrapper[4789]: E1008 15:45:08.785532 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e358a26b-0e6c-4f91-971b-a684fcd7c502" containerName="collect-profiles" Oct 08 15:45:08 crc kubenswrapper[4789]: I1008 15:45:08.785547 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="e358a26b-0e6c-4f91-971b-a684fcd7c502" containerName="collect-profiles" Oct 08 15:45:08 crc kubenswrapper[4789]: I1008 15:45:08.785823 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="e358a26b-0e6c-4f91-971b-a684fcd7c502" containerName="collect-profiles" Oct 08 15:45:08 crc kubenswrapper[4789]: I1008 15:45:08.787728 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:08 crc kubenswrapper[4789]: I1008 15:45:08.801630 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pj669"] Oct 08 15:45:08 crc kubenswrapper[4789]: I1008 15:45:08.985528 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-utilities\") pod \"certified-operators-pj669\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:08 crc kubenswrapper[4789]: I1008 15:45:08.985601 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn4sd\" (UniqueName: \"kubernetes.io/projected/6a4d8721-43af-4733-8fa5-79dada4c35ac-kube-api-access-tn4sd\") pod \"certified-operators-pj669\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:08 crc kubenswrapper[4789]: I1008 15:45:08.985909 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-catalog-content\") pod \"certified-operators-pj669\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:09 crc kubenswrapper[4789]: I1008 15:45:09.091605 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-catalog-content\") pod \"certified-operators-pj669\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:09 crc kubenswrapper[4789]: I1008 15:45:09.091894 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-utilities\") pod \"certified-operators-pj669\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:09 crc kubenswrapper[4789]: I1008 15:45:09.092027 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn4sd\" (UniqueName: \"kubernetes.io/projected/6a4d8721-43af-4733-8fa5-79dada4c35ac-kube-api-access-tn4sd\") pod \"certified-operators-pj669\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:09 crc kubenswrapper[4789]: I1008 15:45:09.092439 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-catalog-content\") pod \"certified-operators-pj669\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:09 crc kubenswrapper[4789]: I1008 15:45:09.092520 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-utilities\") pod \"certified-operators-pj669\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:09 crc kubenswrapper[4789]: I1008 15:45:09.123273 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn4sd\" (UniqueName: \"kubernetes.io/projected/6a4d8721-43af-4733-8fa5-79dada4c35ac-kube-api-access-tn4sd\") pod \"certified-operators-pj669\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:09 crc kubenswrapper[4789]: I1008 15:45:09.414074 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:09 crc kubenswrapper[4789]: I1008 15:45:09.949916 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pj669"] Oct 08 15:45:10 crc kubenswrapper[4789]: I1008 15:45:10.783266 4789 generic.go:334] "Generic (PLEG): container finished" podID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerID="0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf" exitCode=0 Oct 08 15:45:10 crc kubenswrapper[4789]: I1008 15:45:10.783321 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj669" event={"ID":"6a4d8721-43af-4733-8fa5-79dada4c35ac","Type":"ContainerDied","Data":"0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf"} Oct 08 15:45:10 crc kubenswrapper[4789]: I1008 15:45:10.783565 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj669" event={"ID":"6a4d8721-43af-4733-8fa5-79dada4c35ac","Type":"ContainerStarted","Data":"42de870f1999ec2ea1d34d6dc53ce3f818b3d6bce459bdc82bcf0c614de010fd"} Oct 08 15:45:11 crc kubenswrapper[4789]: I1008 15:45:11.799091 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj669" event={"ID":"6a4d8721-43af-4733-8fa5-79dada4c35ac","Type":"ContainerStarted","Data":"8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6"} Oct 08 15:45:12 crc kubenswrapper[4789]: I1008 15:45:12.810034 4789 generic.go:334] "Generic (PLEG): container finished" podID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerID="8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6" exitCode=0 Oct 08 15:45:12 crc kubenswrapper[4789]: I1008 15:45:12.810224 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj669" event={"ID":"6a4d8721-43af-4733-8fa5-79dada4c35ac","Type":"ContainerDied","Data":"8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6"} Oct 08 15:45:13 crc kubenswrapper[4789]: I1008 15:45:13.822951 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj669" event={"ID":"6a4d8721-43af-4733-8fa5-79dada4c35ac","Type":"ContainerStarted","Data":"0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b"} Oct 08 15:45:13 crc kubenswrapper[4789]: I1008 15:45:13.845544 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pj669" podStartSLOduration=3.43137162 podStartE2EDuration="5.845524549s" podCreationTimestamp="2025-10-08 15:45:08 +0000 UTC" firstStartedPulling="2025-10-08 15:45:10.785299969 +0000 UTC m=+6250.692047461" lastFinishedPulling="2025-10-08 15:45:13.199452908 +0000 UTC m=+6253.106200390" observedRunningTime="2025-10-08 15:45:13.839665829 +0000 UTC m=+6253.746413321" watchObservedRunningTime="2025-10-08 15:45:13.845524549 +0000 UTC m=+6253.752272041" Oct 08 15:45:19 crc kubenswrapper[4789]: I1008 15:45:19.414599 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:19 crc kubenswrapper[4789]: I1008 15:45:19.415143 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:19 crc kubenswrapper[4789]: I1008 15:45:19.469968 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:19 crc kubenswrapper[4789]: I1008 15:45:19.956361 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:20 crc kubenswrapper[4789]: I1008 15:45:20.012737 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pj669"] Oct 08 15:45:21 crc kubenswrapper[4789]: I1008 15:45:21.924469 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pj669" podUID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerName="registry-server" containerID="cri-o://0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b" gracePeriod=2 Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.428657 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.575646 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-utilities\") pod \"6a4d8721-43af-4733-8fa5-79dada4c35ac\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.575700 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-catalog-content\") pod \"6a4d8721-43af-4733-8fa5-79dada4c35ac\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.575760 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tn4sd\" (UniqueName: \"kubernetes.io/projected/6a4d8721-43af-4733-8fa5-79dada4c35ac-kube-api-access-tn4sd\") pod \"6a4d8721-43af-4733-8fa5-79dada4c35ac\" (UID: \"6a4d8721-43af-4733-8fa5-79dada4c35ac\") " Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.576624 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-utilities" (OuterVolumeSpecName: "utilities") pod "6a4d8721-43af-4733-8fa5-79dada4c35ac" (UID: "6a4d8721-43af-4733-8fa5-79dada4c35ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.583124 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a4d8721-43af-4733-8fa5-79dada4c35ac-kube-api-access-tn4sd" (OuterVolumeSpecName: "kube-api-access-tn4sd") pod "6a4d8721-43af-4733-8fa5-79dada4c35ac" (UID: "6a4d8721-43af-4733-8fa5-79dada4c35ac"). InnerVolumeSpecName "kube-api-access-tn4sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.694051 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.694111 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tn4sd\" (UniqueName: \"kubernetes.io/projected/6a4d8721-43af-4733-8fa5-79dada4c35ac-kube-api-access-tn4sd\") on node \"crc\" DevicePath \"\"" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.863332 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a4d8721-43af-4733-8fa5-79dada4c35ac" (UID: "6a4d8721-43af-4733-8fa5-79dada4c35ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.899764 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a4d8721-43af-4733-8fa5-79dada4c35ac-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.935010 4789 generic.go:334] "Generic (PLEG): container finished" podID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerID="0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b" exitCode=0 Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.935056 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj669" event={"ID":"6a4d8721-43af-4733-8fa5-79dada4c35ac","Type":"ContainerDied","Data":"0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b"} Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.935061 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pj669" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.935095 4789 scope.go:117] "RemoveContainer" containerID="0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.935085 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pj669" event={"ID":"6a4d8721-43af-4733-8fa5-79dada4c35ac","Type":"ContainerDied","Data":"42de870f1999ec2ea1d34d6dc53ce3f818b3d6bce459bdc82bcf0c614de010fd"} Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.973192 4789 scope.go:117] "RemoveContainer" containerID="8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6" Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.974583 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pj669"] Oct 08 15:45:22 crc kubenswrapper[4789]: I1008 15:45:22.987827 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pj669"] Oct 08 15:45:23 crc kubenswrapper[4789]: I1008 15:45:23.010637 4789 scope.go:117] "RemoveContainer" containerID="0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf" Oct 08 15:45:23 crc kubenswrapper[4789]: I1008 15:45:23.057028 4789 scope.go:117] "RemoveContainer" containerID="0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b" Oct 08 15:45:23 crc kubenswrapper[4789]: E1008 15:45:23.057546 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b\": container with ID starting with 0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b not found: ID does not exist" containerID="0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b" Oct 08 15:45:23 crc kubenswrapper[4789]: I1008 15:45:23.057575 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b"} err="failed to get container status \"0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b\": rpc error: code = NotFound desc = could not find container \"0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b\": container with ID starting with 0dd2b47b64c5ea0dd710fc6fcc7193bbab798a9fd9d03a9daddb18274a19890b not found: ID does not exist" Oct 08 15:45:23 crc kubenswrapper[4789]: I1008 15:45:23.057600 4789 scope.go:117] "RemoveContainer" containerID="8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6" Oct 08 15:45:23 crc kubenswrapper[4789]: E1008 15:45:23.057902 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6\": container with ID starting with 8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6 not found: ID does not exist" containerID="8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6" Oct 08 15:45:23 crc kubenswrapper[4789]: I1008 15:45:23.057946 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6"} err="failed to get container status \"8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6\": rpc error: code = NotFound desc = could not find container \"8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6\": container with ID starting with 8400a4786f01cafd5d538c8821a32bb713593e2d372e91f4c279014122114ff6 not found: ID does not exist" Oct 08 15:45:23 crc kubenswrapper[4789]: I1008 15:45:23.057977 4789 scope.go:117] "RemoveContainer" containerID="0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf" Oct 08 15:45:23 crc kubenswrapper[4789]: E1008 15:45:23.058363 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf\": container with ID starting with 0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf not found: ID does not exist" containerID="0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf" Oct 08 15:45:23 crc kubenswrapper[4789]: I1008 15:45:23.058390 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf"} err="failed to get container status \"0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf\": rpc error: code = NotFound desc = could not find container \"0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf\": container with ID starting with 0878afbf0f91c9a9185c551901d08848a873990e84fe227ad03c8d4d99c4d3cf not found: ID does not exist" Oct 08 15:45:24 crc kubenswrapper[4789]: I1008 15:45:24.741315 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a4d8721-43af-4733-8fa5-79dada4c35ac" path="/var/lib/kubelet/pods/6a4d8721-43af-4733-8fa5-79dada4c35ac/volumes" Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.433278 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.433614 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.433662 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.434568 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bd9daa530e0df0e138a935950ee0c210dd40007a2821f368084c00d54709c8aa"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.434638 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://bd9daa530e0df0e138a935950ee0c210dd40007a2821f368084c00d54709c8aa" gracePeriod=600 Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.702750 4789 scope.go:117] "RemoveContainer" containerID="ddbc84812cf1e0be05faf746d2bab1c1356354e5d74adf6d914da9602e3c6a43" Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.974535 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="bd9daa530e0df0e138a935950ee0c210dd40007a2821f368084c00d54709c8aa" exitCode=0 Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.974608 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"bd9daa530e0df0e138a935950ee0c210dd40007a2821f368084c00d54709c8aa"} Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.974934 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a"} Oct 08 15:45:26 crc kubenswrapper[4789]: I1008 15:45:26.974956 4789 scope.go:117] "RemoveContainer" containerID="1c0d1b50f9936bdf35846feb23b695f677753079fd6909a5027b6f13e5b5524a" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.223568 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ws6vd/must-gather-v68cr"] Oct 08 15:45:36 crc kubenswrapper[4789]: E1008 15:45:36.237537 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerName="extract-utilities" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.237564 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerName="extract-utilities" Oct 08 15:45:36 crc kubenswrapper[4789]: E1008 15:45:36.237588 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerName="extract-content" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.237596 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerName="extract-content" Oct 08 15:45:36 crc kubenswrapper[4789]: E1008 15:45:36.237635 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerName="registry-server" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.237665 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerName="registry-server" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.237975 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a4d8721-43af-4733-8fa5-79dada4c35ac" containerName="registry-server" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.239360 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.243397 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ws6vd"/"kube-root-ca.crt" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.243634 4789 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ws6vd"/"openshift-service-ca.crt" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.244396 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ws6vd/must-gather-v68cr"] Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.300280 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3401a218-11d5-4000-b639-1f2dccc3b80b-must-gather-output\") pod \"must-gather-v68cr\" (UID: \"3401a218-11d5-4000-b639-1f2dccc3b80b\") " pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.300454 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2s2w\" (UniqueName: \"kubernetes.io/projected/3401a218-11d5-4000-b639-1f2dccc3b80b-kube-api-access-h2s2w\") pod \"must-gather-v68cr\" (UID: \"3401a218-11d5-4000-b639-1f2dccc3b80b\") " pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.404015 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2s2w\" (UniqueName: \"kubernetes.io/projected/3401a218-11d5-4000-b639-1f2dccc3b80b-kube-api-access-h2s2w\") pod \"must-gather-v68cr\" (UID: \"3401a218-11d5-4000-b639-1f2dccc3b80b\") " pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.404251 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3401a218-11d5-4000-b639-1f2dccc3b80b-must-gather-output\") pod \"must-gather-v68cr\" (UID: \"3401a218-11d5-4000-b639-1f2dccc3b80b\") " pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.405546 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3401a218-11d5-4000-b639-1f2dccc3b80b-must-gather-output\") pod \"must-gather-v68cr\" (UID: \"3401a218-11d5-4000-b639-1f2dccc3b80b\") " pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.431903 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2s2w\" (UniqueName: \"kubernetes.io/projected/3401a218-11d5-4000-b639-1f2dccc3b80b-kube-api-access-h2s2w\") pod \"must-gather-v68cr\" (UID: \"3401a218-11d5-4000-b639-1f2dccc3b80b\") " pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:45:36 crc kubenswrapper[4789]: I1008 15:45:36.566774 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:45:37 crc kubenswrapper[4789]: I1008 15:45:37.288531 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ws6vd/must-gather-v68cr"] Oct 08 15:45:38 crc kubenswrapper[4789]: I1008 15:45:38.089584 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/must-gather-v68cr" event={"ID":"3401a218-11d5-4000-b639-1f2dccc3b80b","Type":"ContainerStarted","Data":"272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151"} Oct 08 15:45:38 crc kubenswrapper[4789]: I1008 15:45:38.090272 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/must-gather-v68cr" event={"ID":"3401a218-11d5-4000-b639-1f2dccc3b80b","Type":"ContainerStarted","Data":"b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae"} Oct 08 15:45:38 crc kubenswrapper[4789]: I1008 15:45:38.090299 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/must-gather-v68cr" event={"ID":"3401a218-11d5-4000-b639-1f2dccc3b80b","Type":"ContainerStarted","Data":"259578f50283bba02079aac402038beb34105a27ca57c7c33644c420251c9c29"} Oct 08 15:45:38 crc kubenswrapper[4789]: I1008 15:45:38.112161 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ws6vd/must-gather-v68cr" podStartSLOduration=2.112144934 podStartE2EDuration="2.112144934s" podCreationTimestamp="2025-10-08 15:45:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:45:38.103092358 +0000 UTC m=+6278.009839870" watchObservedRunningTime="2025-10-08 15:45:38.112144934 +0000 UTC m=+6278.018892426" Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.301342 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ws6vd/crc-debug-4q7gt"] Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.303500 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.307795 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ws6vd"/"default-dockercfg-zlpdv" Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.317806 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztvjl\" (UniqueName: \"kubernetes.io/projected/ef08392f-e9b2-4696-bbbe-2248382c0fe2-kube-api-access-ztvjl\") pod \"crc-debug-4q7gt\" (UID: \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\") " pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.317896 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef08392f-e9b2-4696-bbbe-2248382c0fe2-host\") pod \"crc-debug-4q7gt\" (UID: \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\") " pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.419895 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztvjl\" (UniqueName: \"kubernetes.io/projected/ef08392f-e9b2-4696-bbbe-2248382c0fe2-kube-api-access-ztvjl\") pod \"crc-debug-4q7gt\" (UID: \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\") " pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.420009 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef08392f-e9b2-4696-bbbe-2248382c0fe2-host\") pod \"crc-debug-4q7gt\" (UID: \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\") " pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.420138 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef08392f-e9b2-4696-bbbe-2248382c0fe2-host\") pod \"crc-debug-4q7gt\" (UID: \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\") " pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.442714 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztvjl\" (UniqueName: \"kubernetes.io/projected/ef08392f-e9b2-4696-bbbe-2248382c0fe2-kube-api-access-ztvjl\") pod \"crc-debug-4q7gt\" (UID: \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\") " pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:45:41 crc kubenswrapper[4789]: I1008 15:45:41.623557 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:45:41 crc kubenswrapper[4789]: W1008 15:45:41.652390 4789 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef08392f_e9b2_4696_bbbe_2248382c0fe2.slice/crio-689072a7867af3e559b77af9b49734e0fc6ab1c4044aeed1cf34887acdcb443c WatchSource:0}: Error finding container 689072a7867af3e559b77af9b49734e0fc6ab1c4044aeed1cf34887acdcb443c: Status 404 returned error can't find the container with id 689072a7867af3e559b77af9b49734e0fc6ab1c4044aeed1cf34887acdcb443c Oct 08 15:45:42 crc kubenswrapper[4789]: I1008 15:45:42.169274 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" event={"ID":"ef08392f-e9b2-4696-bbbe-2248382c0fe2","Type":"ContainerStarted","Data":"1a22986e1795292d8b8e81e68095dc7d01d96a36733c8339a563f054315a04f3"} Oct 08 15:45:42 crc kubenswrapper[4789]: I1008 15:45:42.169605 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" event={"ID":"ef08392f-e9b2-4696-bbbe-2248382c0fe2","Type":"ContainerStarted","Data":"689072a7867af3e559b77af9b49734e0fc6ab1c4044aeed1cf34887acdcb443c"} Oct 08 15:46:59 crc kubenswrapper[4789]: I1008 15:46:59.789559 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-59fddc7b48-jvk6w_1033617f-6aaf-4d50-bd32-fb0eaf6c99b8/barbican-api/0.log" Oct 08 15:46:59 crc kubenswrapper[4789]: I1008 15:46:59.799949 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-59fddc7b48-jvk6w_1033617f-6aaf-4d50-bd32-fb0eaf6c99b8/barbican-api-log/0.log" Oct 08 15:47:00 crc kubenswrapper[4789]: I1008 15:47:00.011218 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5db9877796-wf2x5_01d2c41c-c841-46c8-84a5-d03987977dba/barbican-keystone-listener/0.log" Oct 08 15:47:00 crc kubenswrapper[4789]: I1008 15:47:00.148189 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5db9877796-wf2x5_01d2c41c-c841-46c8-84a5-d03987977dba/barbican-keystone-listener-log/0.log" Oct 08 15:47:00 crc kubenswrapper[4789]: I1008 15:47:00.290260 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-56d947cc6f-sgvkl_3028d0aa-cb65-4b14-aebb-7b9cb4910eed/barbican-worker/0.log" Oct 08 15:47:00 crc kubenswrapper[4789]: I1008 15:47:00.389344 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-56d947cc6f-sgvkl_3028d0aa-cb65-4b14-aebb-7b9cb4910eed/barbican-worker-log/0.log" Oct 08 15:47:00 crc kubenswrapper[4789]: I1008 15:47:00.488459 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gjrvj_5f34622a-a071-45a1-b7fd-75fa8fbb70fe/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:00 crc kubenswrapper[4789]: I1008 15:47:00.960236 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3152634e-3d68-4544-ad93-2999a2273f3a/ceilometer-notification-agent/0.log" Oct 08 15:47:00 crc kubenswrapper[4789]: I1008 15:47:00.997842 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3152634e-3d68-4544-ad93-2999a2273f3a/ceilometer-central-agent/0.log" Oct 08 15:47:01 crc kubenswrapper[4789]: I1008 15:47:01.103151 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3152634e-3d68-4544-ad93-2999a2273f3a/proxy-httpd/0.log" Oct 08 15:47:01 crc kubenswrapper[4789]: I1008 15:47:01.175786 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3152634e-3d68-4544-ad93-2999a2273f3a/sg-core/0.log" Oct 08 15:47:01 crc kubenswrapper[4789]: I1008 15:47:01.377770 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a72d40d2-86d8-406f-9f5c-adec376cc947/cinder-api-log/0.log" Oct 08 15:47:01 crc kubenswrapper[4789]: I1008 15:47:01.846133 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_6b3acb2d-0c53-40a4-8857-8e07cf54c530/probe/0.log" Oct 08 15:47:02 crc kubenswrapper[4789]: I1008 15:47:02.049449 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_6b3acb2d-0c53-40a4-8857-8e07cf54c530/cinder-backup/0.log" Oct 08 15:47:02 crc kubenswrapper[4789]: I1008 15:47:02.072552 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a72d40d2-86d8-406f-9f5c-adec376cc947/cinder-api/0.log" Oct 08 15:47:02 crc kubenswrapper[4789]: I1008 15:47:02.376742 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4717d784-db56-4840-8b93-9be19efb814b/cinder-scheduler/0.log" Oct 08 15:47:02 crc kubenswrapper[4789]: I1008 15:47:02.385420 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4717d784-db56-4840-8b93-9be19efb814b/probe/0.log" Oct 08 15:47:02 crc kubenswrapper[4789]: I1008 15:47:02.640899 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_7b5501af-92e0-4c4a-badd-05089ae1221b/probe/0.log" Oct 08 15:47:02 crc kubenswrapper[4789]: I1008 15:47:02.797758 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_7b5501af-92e0-4c4a-badd-05089ae1221b/cinder-volume/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.079530 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_e0006437-3cf9-437c-a3bd-a12f6ef84aae/probe/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.148752 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-tbmn9_48dac93b-b8ea-4016-9a48-36fd0c594bf4/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.171746 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_e0006437-3cf9-437c-a3bd-a12f6ef84aae/cinder-volume/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.312849 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-n6wqd_6f2ef14f-0023-4c8d-8153-5dba30e84c17/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.421689 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-znv9w_167cf260-5310-4c48-9911-b5257c1ba256/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.549878 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-65bf458dd9-l6k8r_dbf93231-9ffa-4434-a4c5-3bf97ee66e71/init/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.667604 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-65bf458dd9-l6k8r_dbf93231-9ffa-4434-a4c5-3bf97ee66e71/init/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.778629 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-zp7sd_617364bf-6612-48bf-bfb4-0eeaf9ff2c60/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.888010 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-65bf458dd9-l6k8r_dbf93231-9ffa-4434-a4c5-3bf97ee66e71/dnsmasq-dns/0.log" Oct 08 15:47:03 crc kubenswrapper[4789]: I1008 15:47:03.985690 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1fbecfec-ecd2-458e-b04f-e715a9ec4186/glance-httpd/0.log" Oct 08 15:47:04 crc kubenswrapper[4789]: I1008 15:47:04.010996 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1fbecfec-ecd2-458e-b04f-e715a9ec4186/glance-log/0.log" Oct 08 15:47:04 crc kubenswrapper[4789]: I1008 15:47:04.152451 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a8fbeb79-b731-40fe-867d-1df6c03db681/glance-httpd/0.log" Oct 08 15:47:04 crc kubenswrapper[4789]: I1008 15:47:04.181505 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a8fbeb79-b731-40fe-867d-1df6c03db681/glance-log/0.log" Oct 08 15:47:04 crc kubenswrapper[4789]: I1008 15:47:04.424976 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6c7bdc9dd-bg9vf_f9289b28-1fb4-47f8-aec3-3d9d616837d3/horizon/0.log" Oct 08 15:47:04 crc kubenswrapper[4789]: I1008 15:47:04.543505 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-pjpkm_92d5b86d-9984-45c5-8287-7b3094b4332f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:04 crc kubenswrapper[4789]: I1008 15:47:04.805600 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-xpffg_5a8f80d9-3b33-479c-9d6a-77867520218d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:05 crc kubenswrapper[4789]: I1008 15:47:05.154159 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6c7bdc9dd-bg9vf_f9289b28-1fb4-47f8-aec3-3d9d616837d3/horizon-log/0.log" Oct 08 15:47:05 crc kubenswrapper[4789]: I1008 15:47:05.333718 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29332261-6nq76_caa89bd3-cdc8-412a-82f0-1289cf0028c1/keystone-cron/0.log" Oct 08 15:47:05 crc kubenswrapper[4789]: I1008 15:47:05.512904 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_08e907c7-3b60-4d50-977c-32d05a3ce80f/kube-state-metrics/0.log" Oct 08 15:47:05 crc kubenswrapper[4789]: I1008 15:47:05.526629 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c9c77979c-fl8sh_5eeed1f7-b400-4c25-851d-43a1f036fa8b/keystone-api/0.log" Oct 08 15:47:05 crc kubenswrapper[4789]: I1008 15:47:05.727530 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-kv8zg_2066f52b-bb99-4a61-98e4-63f53b8ec7c7/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:06 crc kubenswrapper[4789]: I1008 15:47:06.285865 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7dcb5d4fcc-q9spz_1532f955-2e73-4a9a-b76a-507c2c545db0/neutron-httpd/0.log" Oct 08 15:47:06 crc kubenswrapper[4789]: I1008 15:47:06.291571 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7dcb5d4fcc-q9spz_1532f955-2e73-4a9a-b76a-507c2c545db0/neutron-api/0.log" Oct 08 15:47:06 crc kubenswrapper[4789]: I1008 15:47:06.439509 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-2jbdt_7f2ef2f7-aa9f-4767-a032-e461dde85655/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:07 crc kubenswrapper[4789]: I1008 15:47:07.408085 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_de1fa794-a227-446e-a3b0-2b1c120e4ecf/nova-cell0-conductor-conductor/0.log" Oct 08 15:47:08 crc kubenswrapper[4789]: I1008 15:47:08.088631 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_af59a4db-7cb6-4e1f-8381-43e0f13e5875/nova-cell1-conductor-conductor/0.log" Oct 08 15:47:08 crc kubenswrapper[4789]: I1008 15:47:08.227682 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c/nova-api-log/0.log" Oct 08 15:47:08 crc kubenswrapper[4789]: I1008 15:47:08.452952 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9a9aa57f-9cd9-4b28-8f7a-c9a0cecb0a1c/nova-api-api/0.log" Oct 08 15:47:08 crc kubenswrapper[4789]: I1008 15:47:08.580293 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_42136e0e-1d9e-44ea-b9fc-3e7e313fe107/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 15:47:08 crc kubenswrapper[4789]: I1008 15:47:08.774601 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-gbfjb_b8ce1628-fa93-4910-af5a-ebce3d8ecc30/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:08 crc kubenswrapper[4789]: I1008 15:47:08.936061 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9255b8e1-bd22-43cf-b223-87021f091449/nova-metadata-log/0.log" Oct 08 15:47:09 crc kubenswrapper[4789]: I1008 15:47:09.652232 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ad6bc841-5112-4ce6-a8ae-926e8b7d8043/nova-scheduler-scheduler/0.log" Oct 08 15:47:09 crc kubenswrapper[4789]: I1008 15:47:09.749294 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_32ceb31c-1751-4a04-a91a-b317f3ba3d78/mysql-bootstrap/0.log" Oct 08 15:47:09 crc kubenswrapper[4789]: I1008 15:47:09.923177 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_32ceb31c-1751-4a04-a91a-b317f3ba3d78/mysql-bootstrap/0.log" Oct 08 15:47:09 crc kubenswrapper[4789]: I1008 15:47:09.960537 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_32ceb31c-1751-4a04-a91a-b317f3ba3d78/galera/0.log" Oct 08 15:47:10 crc kubenswrapper[4789]: I1008 15:47:10.166220 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_505cf346-64b1-4de8-8cf6-fca4ddd34cd6/mysql-bootstrap/0.log" Oct 08 15:47:10 crc kubenswrapper[4789]: I1008 15:47:10.500409 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_505cf346-64b1-4de8-8cf6-fca4ddd34cd6/mysql-bootstrap/0.log" Oct 08 15:47:10 crc kubenswrapper[4789]: I1008 15:47:10.525153 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_505cf346-64b1-4de8-8cf6-fca4ddd34cd6/galera/0.log" Oct 08 15:47:10 crc kubenswrapper[4789]: I1008 15:47:10.767902 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2ba79b7a-39cd-4f22-8094-672ef7b58e76/openstackclient/0.log" Oct 08 15:47:11 crc kubenswrapper[4789]: I1008 15:47:11.016425 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nrhvq_23e0c1e2-0a8f-4f65-973a-a2bfe4b8f718/openstack-network-exporter/0.log" Oct 08 15:47:11 crc kubenswrapper[4789]: I1008 15:47:11.280415 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tkv8v_5e2285a3-7296-490c-8649-adeae9a0cabc/ovsdb-server-init/0.log" Oct 08 15:47:11 crc kubenswrapper[4789]: I1008 15:47:11.481784 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tkv8v_5e2285a3-7296-490c-8649-adeae9a0cabc/ovsdb-server-init/0.log" Oct 08 15:47:11 crc kubenswrapper[4789]: I1008 15:47:11.647879 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tkv8v_5e2285a3-7296-490c-8649-adeae9a0cabc/ovsdb-server/0.log" Oct 08 15:47:11 crc kubenswrapper[4789]: I1008 15:47:11.735981 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9255b8e1-bd22-43cf-b223-87021f091449/nova-metadata-metadata/0.log" Oct 08 15:47:11 crc kubenswrapper[4789]: I1008 15:47:11.841065 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tkv8v_5e2285a3-7296-490c-8649-adeae9a0cabc/ovs-vswitchd/0.log" Oct 08 15:47:11 crc kubenswrapper[4789]: I1008 15:47:11.957133 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zrdhb_91ddb1cc-17c9-4f38-ac8a-d78d10798178/ovn-controller/0.log" Oct 08 15:47:12 crc kubenswrapper[4789]: I1008 15:47:12.202015 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-4djc2_516d476a-1a73-4e62-a06f-d9eb90541500/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:12 crc kubenswrapper[4789]: I1008 15:47:12.290561 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fd23ae36-37d1-458e-ab17-0f0db57e57d3/openstack-network-exporter/0.log" Oct 08 15:47:12 crc kubenswrapper[4789]: I1008 15:47:12.411492 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fd23ae36-37d1-458e-ab17-0f0db57e57d3/ovn-northd/0.log" Oct 08 15:47:12 crc kubenswrapper[4789]: I1008 15:47:12.490836 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a1a9507e-04fe-4746-a31b-5ad097c48602/openstack-network-exporter/0.log" Oct 08 15:47:12 crc kubenswrapper[4789]: I1008 15:47:12.642501 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a1a9507e-04fe-4746-a31b-5ad097c48602/ovsdbserver-nb/0.log" Oct 08 15:47:12 crc kubenswrapper[4789]: I1008 15:47:12.750441 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_70dc7aa5-b81d-4150-b1dd-57ec365e495c/openstack-network-exporter/0.log" Oct 08 15:47:12 crc kubenswrapper[4789]: I1008 15:47:12.963966 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_70dc7aa5-b81d-4150-b1dd-57ec365e495c/ovsdbserver-sb/0.log" Oct 08 15:47:13 crc kubenswrapper[4789]: I1008 15:47:13.165457 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6c99f9dffd-c2zmb_c79d9b2e-fb14-4d73-a6f0-d8b0684001ac/placement-api/0.log" Oct 08 15:47:13 crc kubenswrapper[4789]: I1008 15:47:13.367801 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/init-config-reloader/0.log" Oct 08 15:47:13 crc kubenswrapper[4789]: I1008 15:47:13.447881 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6c99f9dffd-c2zmb_c79d9b2e-fb14-4d73-a6f0-d8b0684001ac/placement-log/0.log" Oct 08 15:47:13 crc kubenswrapper[4789]: I1008 15:47:13.626200 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/init-config-reloader/0.log" Oct 08 15:47:13 crc kubenswrapper[4789]: I1008 15:47:13.626557 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/config-reloader/0.log" Oct 08 15:47:13 crc kubenswrapper[4789]: I1008 15:47:13.689903 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/prometheus/0.log" Oct 08 15:47:13 crc kubenswrapper[4789]: I1008 15:47:13.816315 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_441b4491-1181-4811-af7f-07e8e672352f/thanos-sidecar/0.log" Oct 08 15:47:13 crc kubenswrapper[4789]: I1008 15:47:13.929837 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c4758067-9242-48d2-a066-e007669ef85c/setup-container/0.log" Oct 08 15:47:14 crc kubenswrapper[4789]: I1008 15:47:14.394431 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c4758067-9242-48d2-a066-e007669ef85c/rabbitmq/0.log" Oct 08 15:47:14 crc kubenswrapper[4789]: I1008 15:47:14.445064 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c4758067-9242-48d2-a066-e007669ef85c/setup-container/0.log" Oct 08 15:47:14 crc kubenswrapper[4789]: I1008 15:47:14.566926 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_55fc0a2c-48c9-4508-ae9d-5350f3298621/setup-container/0.log" Oct 08 15:47:14 crc kubenswrapper[4789]: I1008 15:47:14.799186 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_55fc0a2c-48c9-4508-ae9d-5350f3298621/rabbitmq/0.log" Oct 08 15:47:14 crc kubenswrapper[4789]: I1008 15:47:14.803576 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_55fc0a2c-48c9-4508-ae9d-5350f3298621/setup-container/0.log" Oct 08 15:47:15 crc kubenswrapper[4789]: I1008 15:47:15.033515 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_698f294c-233d-45d0-a47b-f3580b23b955/setup-container/0.log" Oct 08 15:47:15 crc kubenswrapper[4789]: I1008 15:47:15.189867 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_698f294c-233d-45d0-a47b-f3580b23b955/setup-container/0.log" Oct 08 15:47:15 crc kubenswrapper[4789]: I1008 15:47:15.263809 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_698f294c-233d-45d0-a47b-f3580b23b955/rabbitmq/0.log" Oct 08 15:47:15 crc kubenswrapper[4789]: I1008 15:47:15.443329 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-6mgmg_f2575ace-ab3d-475e-b049-a2d2614104bf/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:15 crc kubenswrapper[4789]: I1008 15:47:15.581327 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-785jj_c5a284ba-b502-45cc-b9b6-f460908bee2f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:15 crc kubenswrapper[4789]: I1008 15:47:15.764880 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-p59gr_484052d0-6aeb-4f53-bad2-d6228aaadcf3/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:15 crc kubenswrapper[4789]: I1008 15:47:15.988422 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-xz5th_1ef78bfb-47eb-40df-93d6-7b6070c33f5f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:16 crc kubenswrapper[4789]: I1008 15:47:16.083264 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-29nmw_ad2db88c-41b1-4feb-b579-96bb456cea2a/ssh-known-hosts-edpm-deployment/0.log" Oct 08 15:47:16 crc kubenswrapper[4789]: I1008 15:47:16.329614 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7bf8478fbc-fpmxn_ba045b83-d6ec-4760-98dd-b330dca8ed99/proxy-server/0.log" Oct 08 15:47:16 crc kubenswrapper[4789]: I1008 15:47:16.535922 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7bf8478fbc-fpmxn_ba045b83-d6ec-4760-98dd-b330dca8ed99/proxy-httpd/0.log" Oct 08 15:47:16 crc kubenswrapper[4789]: I1008 15:47:16.562673 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-pr6sr_a7b17c27-5518-480a-875f-091ae1dd3606/swift-ring-rebalance/0.log" Oct 08 15:47:16 crc kubenswrapper[4789]: I1008 15:47:16.770832 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/account-auditor/0.log" Oct 08 15:47:16 crc kubenswrapper[4789]: I1008 15:47:16.788882 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/account-reaper/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.025181 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/account-server/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.039867 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/account-replicator/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.060160 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/container-auditor/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.229179 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/container-server/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.304942 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/container-updater/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.327706 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/container-replicator/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.503360 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-auditor/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.504375 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-expirer/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.538775 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-replicator/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.707018 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-server/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.740622 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/object-updater/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.780750 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/rsync/0.log" Oct 08 15:47:17 crc kubenswrapper[4789]: I1008 15:47:17.970941 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0462c12d-8df3-4734-9c2e-a925c9c06b9e/swift-recon-cron/0.log" Oct 08 15:47:18 crc kubenswrapper[4789]: I1008 15:47:18.023590 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-x5pkk_c9031f68-f885-456b-a1ea-ce7742e0eccc/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:18 crc kubenswrapper[4789]: I1008 15:47:18.287268 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5622aa64-be18-4420-8c84-b76fe03c1346/tempest-tests-tempest-tests-runner/0.log" Oct 08 15:47:18 crc kubenswrapper[4789]: I1008 15:47:18.472663 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_3d01bf45-7cf8-44f1-a306-8ed97d937dfe/test-operator-logs-container/0.log" Oct 08 15:47:18 crc kubenswrapper[4789]: I1008 15:47:18.561732 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-6xkvj_04be0498-514b-45df-9ff0-7eae07586037/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 15:47:19 crc kubenswrapper[4789]: I1008 15:47:19.827557 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_d84335e3-9c9e-473f-be0a-f1f4535ae6b5/watcher-applier/0.log" Oct 08 15:47:19 crc kubenswrapper[4789]: I1008 15:47:19.888599 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_4bdbb9f8-14b6-4f34-b818-a9deb84247cf/watcher-api-log/0.log" Oct 08 15:47:23 crc kubenswrapper[4789]: I1008 15:47:23.710174 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_4c9e2e8c-9df6-4340-af61-2778209a54dd/watcher-decision-engine/0.log" Oct 08 15:47:25 crc kubenswrapper[4789]: I1008 15:47:25.059196 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_4bdbb9f8-14b6-4f34-b818-a9deb84247cf/watcher-api/0.log" Oct 08 15:47:26 crc kubenswrapper[4789]: I1008 15:47:26.432288 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:47:26 crc kubenswrapper[4789]: I1008 15:47:26.432639 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:47:26 crc kubenswrapper[4789]: I1008 15:47:26.825433 4789 scope.go:117] "RemoveContainer" containerID="36b4f9eece173562ab3740c21d7afdcdc43d1bcb1de8b6a728eb2cb99185a819" Oct 08 15:47:37 crc kubenswrapper[4789]: I1008 15:47:37.031874 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_78e021d8-ac60-4092-9e49-bbb3fff88aa7/memcached/0.log" Oct 08 15:47:54 crc kubenswrapper[4789]: I1008 15:47:54.526414 4789 generic.go:334] "Generic (PLEG): container finished" podID="ef08392f-e9b2-4696-bbbe-2248382c0fe2" containerID="1a22986e1795292d8b8e81e68095dc7d01d96a36733c8339a563f054315a04f3" exitCode=0 Oct 08 15:47:54 crc kubenswrapper[4789]: I1008 15:47:54.526536 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" event={"ID":"ef08392f-e9b2-4696-bbbe-2248382c0fe2","Type":"ContainerDied","Data":"1a22986e1795292d8b8e81e68095dc7d01d96a36733c8339a563f054315a04f3"} Oct 08 15:47:55 crc kubenswrapper[4789]: I1008 15:47:55.669491 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:47:55 crc kubenswrapper[4789]: I1008 15:47:55.704383 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ws6vd/crc-debug-4q7gt"] Oct 08 15:47:55 crc kubenswrapper[4789]: I1008 15:47:55.712803 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ws6vd/crc-debug-4q7gt"] Oct 08 15:47:55 crc kubenswrapper[4789]: I1008 15:47:55.768481 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztvjl\" (UniqueName: \"kubernetes.io/projected/ef08392f-e9b2-4696-bbbe-2248382c0fe2-kube-api-access-ztvjl\") pod \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\" (UID: \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\") " Oct 08 15:47:55 crc kubenswrapper[4789]: I1008 15:47:55.768522 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef08392f-e9b2-4696-bbbe-2248382c0fe2-host\") pod \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\" (UID: \"ef08392f-e9b2-4696-bbbe-2248382c0fe2\") " Oct 08 15:47:55 crc kubenswrapper[4789]: I1008 15:47:55.768722 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef08392f-e9b2-4696-bbbe-2248382c0fe2-host" (OuterVolumeSpecName: "host") pod "ef08392f-e9b2-4696-bbbe-2248382c0fe2" (UID: "ef08392f-e9b2-4696-bbbe-2248382c0fe2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:47:55 crc kubenswrapper[4789]: I1008 15:47:55.769097 4789 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef08392f-e9b2-4696-bbbe-2248382c0fe2-host\") on node \"crc\" DevicePath \"\"" Oct 08 15:47:55 crc kubenswrapper[4789]: I1008 15:47:55.775188 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef08392f-e9b2-4696-bbbe-2248382c0fe2-kube-api-access-ztvjl" (OuterVolumeSpecName: "kube-api-access-ztvjl") pod "ef08392f-e9b2-4696-bbbe-2248382c0fe2" (UID: "ef08392f-e9b2-4696-bbbe-2248382c0fe2"). InnerVolumeSpecName "kube-api-access-ztvjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:47:55 crc kubenswrapper[4789]: I1008 15:47:55.871032 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztvjl\" (UniqueName: \"kubernetes.io/projected/ef08392f-e9b2-4696-bbbe-2248382c0fe2-kube-api-access-ztvjl\") on node \"crc\" DevicePath \"\"" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.432711 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.432790 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.549402 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="689072a7867af3e559b77af9b49734e0fc6ab1c4044aeed1cf34887acdcb443c" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.549500 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-4q7gt" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.741461 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef08392f-e9b2-4696-bbbe-2248382c0fe2" path="/var/lib/kubelet/pods/ef08392f-e9b2-4696-bbbe-2248382c0fe2/volumes" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.874038 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ws6vd/crc-debug-zw44m"] Oct 08 15:47:56 crc kubenswrapper[4789]: E1008 15:47:56.875249 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef08392f-e9b2-4696-bbbe-2248382c0fe2" containerName="container-00" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.875424 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef08392f-e9b2-4696-bbbe-2248382c0fe2" containerName="container-00" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.876038 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef08392f-e9b2-4696-bbbe-2248382c0fe2" containerName="container-00" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.877341 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.881419 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ws6vd"/"default-dockercfg-zlpdv" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.995481 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4nbv\" (UniqueName: \"kubernetes.io/projected/823ad436-cbf6-45fc-8b55-246f0b9a387b-kube-api-access-t4nbv\") pod \"crc-debug-zw44m\" (UID: \"823ad436-cbf6-45fc-8b55-246f0b9a387b\") " pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:47:56 crc kubenswrapper[4789]: I1008 15:47:56.995618 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823ad436-cbf6-45fc-8b55-246f0b9a387b-host\") pod \"crc-debug-zw44m\" (UID: \"823ad436-cbf6-45fc-8b55-246f0b9a387b\") " pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:47:57 crc kubenswrapper[4789]: I1008 15:47:57.097640 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4nbv\" (UniqueName: \"kubernetes.io/projected/823ad436-cbf6-45fc-8b55-246f0b9a387b-kube-api-access-t4nbv\") pod \"crc-debug-zw44m\" (UID: \"823ad436-cbf6-45fc-8b55-246f0b9a387b\") " pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:47:57 crc kubenswrapper[4789]: I1008 15:47:57.097797 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823ad436-cbf6-45fc-8b55-246f0b9a387b-host\") pod \"crc-debug-zw44m\" (UID: \"823ad436-cbf6-45fc-8b55-246f0b9a387b\") " pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:47:57 crc kubenswrapper[4789]: I1008 15:47:57.098123 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823ad436-cbf6-45fc-8b55-246f0b9a387b-host\") pod \"crc-debug-zw44m\" (UID: \"823ad436-cbf6-45fc-8b55-246f0b9a387b\") " pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:47:57 crc kubenswrapper[4789]: I1008 15:47:57.119946 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4nbv\" (UniqueName: \"kubernetes.io/projected/823ad436-cbf6-45fc-8b55-246f0b9a387b-kube-api-access-t4nbv\") pod \"crc-debug-zw44m\" (UID: \"823ad436-cbf6-45fc-8b55-246f0b9a387b\") " pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:47:57 crc kubenswrapper[4789]: I1008 15:47:57.196574 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:47:57 crc kubenswrapper[4789]: I1008 15:47:57.562566 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/crc-debug-zw44m" event={"ID":"823ad436-cbf6-45fc-8b55-246f0b9a387b","Type":"ContainerStarted","Data":"388f57656d77819dcb55d723c361e9a49b2469e35dbdc2157e2de9cf48d8af5d"} Oct 08 15:47:57 crc kubenswrapper[4789]: I1008 15:47:57.562928 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/crc-debug-zw44m" event={"ID":"823ad436-cbf6-45fc-8b55-246f0b9a387b","Type":"ContainerStarted","Data":"e254db06e97c948565833b04f687f4c494acad21342c600ef735e5706744c344"} Oct 08 15:47:57 crc kubenswrapper[4789]: I1008 15:47:57.587120 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ws6vd/crc-debug-zw44m" podStartSLOduration=1.5870891249999999 podStartE2EDuration="1.587089125s" podCreationTimestamp="2025-10-08 15:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 15:47:57.580560686 +0000 UTC m=+6417.487308178" watchObservedRunningTime="2025-10-08 15:47:57.587089125 +0000 UTC m=+6417.493836637" Oct 08 15:47:58 crc kubenswrapper[4789]: I1008 15:47:58.573464 4789 generic.go:334] "Generic (PLEG): container finished" podID="823ad436-cbf6-45fc-8b55-246f0b9a387b" containerID="388f57656d77819dcb55d723c361e9a49b2469e35dbdc2157e2de9cf48d8af5d" exitCode=0 Oct 08 15:47:58 crc kubenswrapper[4789]: I1008 15:47:58.574108 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/crc-debug-zw44m" event={"ID":"823ad436-cbf6-45fc-8b55-246f0b9a387b","Type":"ContainerDied","Data":"388f57656d77819dcb55d723c361e9a49b2469e35dbdc2157e2de9cf48d8af5d"} Oct 08 15:47:59 crc kubenswrapper[4789]: I1008 15:47:59.712035 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:47:59 crc kubenswrapper[4789]: I1008 15:47:59.749995 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823ad436-cbf6-45fc-8b55-246f0b9a387b-host\") pod \"823ad436-cbf6-45fc-8b55-246f0b9a387b\" (UID: \"823ad436-cbf6-45fc-8b55-246f0b9a387b\") " Oct 08 15:47:59 crc kubenswrapper[4789]: I1008 15:47:59.750045 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/823ad436-cbf6-45fc-8b55-246f0b9a387b-host" (OuterVolumeSpecName: "host") pod "823ad436-cbf6-45fc-8b55-246f0b9a387b" (UID: "823ad436-cbf6-45fc-8b55-246f0b9a387b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:47:59 crc kubenswrapper[4789]: I1008 15:47:59.750576 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4nbv\" (UniqueName: \"kubernetes.io/projected/823ad436-cbf6-45fc-8b55-246f0b9a387b-kube-api-access-t4nbv\") pod \"823ad436-cbf6-45fc-8b55-246f0b9a387b\" (UID: \"823ad436-cbf6-45fc-8b55-246f0b9a387b\") " Oct 08 15:47:59 crc kubenswrapper[4789]: I1008 15:47:59.751315 4789 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823ad436-cbf6-45fc-8b55-246f0b9a387b-host\") on node \"crc\" DevicePath \"\"" Oct 08 15:47:59 crc kubenswrapper[4789]: I1008 15:47:59.756636 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/823ad436-cbf6-45fc-8b55-246f0b9a387b-kube-api-access-t4nbv" (OuterVolumeSpecName: "kube-api-access-t4nbv") pod "823ad436-cbf6-45fc-8b55-246f0b9a387b" (UID: "823ad436-cbf6-45fc-8b55-246f0b9a387b"). InnerVolumeSpecName "kube-api-access-t4nbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:47:59 crc kubenswrapper[4789]: I1008 15:47:59.852965 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4nbv\" (UniqueName: \"kubernetes.io/projected/823ad436-cbf6-45fc-8b55-246f0b9a387b-kube-api-access-t4nbv\") on node \"crc\" DevicePath \"\"" Oct 08 15:48:00 crc kubenswrapper[4789]: I1008 15:48:00.592313 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/crc-debug-zw44m" event={"ID":"823ad436-cbf6-45fc-8b55-246f0b9a387b","Type":"ContainerDied","Data":"e254db06e97c948565833b04f687f4c494acad21342c600ef735e5706744c344"} Oct 08 15:48:00 crc kubenswrapper[4789]: I1008 15:48:00.592648 4789 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e254db06e97c948565833b04f687f4c494acad21342c600ef735e5706744c344" Oct 08 15:48:00 crc kubenswrapper[4789]: I1008 15:48:00.592361 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-zw44m" Oct 08 15:48:07 crc kubenswrapper[4789]: I1008 15:48:07.158501 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ws6vd/crc-debug-zw44m"] Oct 08 15:48:07 crc kubenswrapper[4789]: I1008 15:48:07.166796 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ws6vd/crc-debug-zw44m"] Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.325532 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ws6vd/crc-debug-z9274"] Oct 08 15:48:08 crc kubenswrapper[4789]: E1008 15:48:08.326619 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="823ad436-cbf6-45fc-8b55-246f0b9a387b" containerName="container-00" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.326635 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="823ad436-cbf6-45fc-8b55-246f0b9a387b" containerName="container-00" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.326918 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="823ad436-cbf6-45fc-8b55-246f0b9a387b" containerName="container-00" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.327977 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.330548 4789 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ws6vd"/"default-dockercfg-zlpdv" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.438771 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss29w\" (UniqueName: \"kubernetes.io/projected/a648d189-74dd-4364-b047-c9e535d0d626-kube-api-access-ss29w\") pod \"crc-debug-z9274\" (UID: \"a648d189-74dd-4364-b047-c9e535d0d626\") " pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.439326 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a648d189-74dd-4364-b047-c9e535d0d626-host\") pod \"crc-debug-z9274\" (UID: \"a648d189-74dd-4364-b047-c9e535d0d626\") " pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.540807 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss29w\" (UniqueName: \"kubernetes.io/projected/a648d189-74dd-4364-b047-c9e535d0d626-kube-api-access-ss29w\") pod \"crc-debug-z9274\" (UID: \"a648d189-74dd-4364-b047-c9e535d0d626\") " pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.540972 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a648d189-74dd-4364-b047-c9e535d0d626-host\") pod \"crc-debug-z9274\" (UID: \"a648d189-74dd-4364-b047-c9e535d0d626\") " pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.541086 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a648d189-74dd-4364-b047-c9e535d0d626-host\") pod \"crc-debug-z9274\" (UID: \"a648d189-74dd-4364-b047-c9e535d0d626\") " pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.560679 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss29w\" (UniqueName: \"kubernetes.io/projected/a648d189-74dd-4364-b047-c9e535d0d626-kube-api-access-ss29w\") pod \"crc-debug-z9274\" (UID: \"a648d189-74dd-4364-b047-c9e535d0d626\") " pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.653124 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:08 crc kubenswrapper[4789]: I1008 15:48:08.739982 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="823ad436-cbf6-45fc-8b55-246f0b9a387b" path="/var/lib/kubelet/pods/823ad436-cbf6-45fc-8b55-246f0b9a387b/volumes" Oct 08 15:48:09 crc kubenswrapper[4789]: I1008 15:48:09.677906 4789 generic.go:334] "Generic (PLEG): container finished" podID="a648d189-74dd-4364-b047-c9e535d0d626" containerID="110f1bc621a97b9f6932174a4f644e421fd88a15b456d790bb790c2788a2cd75" exitCode=0 Oct 08 15:48:09 crc kubenswrapper[4789]: I1008 15:48:09.677981 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/crc-debug-z9274" event={"ID":"a648d189-74dd-4364-b047-c9e535d0d626","Type":"ContainerDied","Data":"110f1bc621a97b9f6932174a4f644e421fd88a15b456d790bb790c2788a2cd75"} Oct 08 15:48:09 crc kubenswrapper[4789]: I1008 15:48:09.678038 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/crc-debug-z9274" event={"ID":"a648d189-74dd-4364-b047-c9e535d0d626","Type":"ContainerStarted","Data":"404dd2ba64dd3a940dd172dff13d5ddf49444969e98cff803df7937120ea2442"} Oct 08 15:48:09 crc kubenswrapper[4789]: I1008 15:48:09.720483 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ws6vd/crc-debug-z9274"] Oct 08 15:48:09 crc kubenswrapper[4789]: I1008 15:48:09.728557 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ws6vd/crc-debug-z9274"] Oct 08 15:48:10 crc kubenswrapper[4789]: I1008 15:48:10.809225 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:10 crc kubenswrapper[4789]: I1008 15:48:10.891121 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a648d189-74dd-4364-b047-c9e535d0d626-host\") pod \"a648d189-74dd-4364-b047-c9e535d0d626\" (UID: \"a648d189-74dd-4364-b047-c9e535d0d626\") " Oct 08 15:48:10 crc kubenswrapper[4789]: I1008 15:48:10.891278 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a648d189-74dd-4364-b047-c9e535d0d626-host" (OuterVolumeSpecName: "host") pod "a648d189-74dd-4364-b047-c9e535d0d626" (UID: "a648d189-74dd-4364-b047-c9e535d0d626"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 15:48:10 crc kubenswrapper[4789]: I1008 15:48:10.891794 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss29w\" (UniqueName: \"kubernetes.io/projected/a648d189-74dd-4364-b047-c9e535d0d626-kube-api-access-ss29w\") pod \"a648d189-74dd-4364-b047-c9e535d0d626\" (UID: \"a648d189-74dd-4364-b047-c9e535d0d626\") " Oct 08 15:48:10 crc kubenswrapper[4789]: I1008 15:48:10.893734 4789 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a648d189-74dd-4364-b047-c9e535d0d626-host\") on node \"crc\" DevicePath \"\"" Oct 08 15:48:10 crc kubenswrapper[4789]: I1008 15:48:10.909122 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a648d189-74dd-4364-b047-c9e535d0d626-kube-api-access-ss29w" (OuterVolumeSpecName: "kube-api-access-ss29w") pod "a648d189-74dd-4364-b047-c9e535d0d626" (UID: "a648d189-74dd-4364-b047-c9e535d0d626"). InnerVolumeSpecName "kube-api-access-ss29w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:48:10 crc kubenswrapper[4789]: I1008 15:48:10.995813 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss29w\" (UniqueName: \"kubernetes.io/projected/a648d189-74dd-4364-b047-c9e535d0d626-kube-api-access-ss29w\") on node \"crc\" DevicePath \"\"" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.290181 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/util/0.log" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.452302 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/util/0.log" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.497864 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/pull/0.log" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.502617 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/pull/0.log" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.702705 4789 scope.go:117] "RemoveContainer" containerID="110f1bc621a97b9f6932174a4f644e421fd88a15b456d790bb790c2788a2cd75" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.702777 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/crc-debug-z9274" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.715357 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/pull/0.log" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.748040 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/util/0.log" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.759107 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_96547d1085696b28a94ff813ab1cf03e0043cc7380c31b115c6d4db9dcb4xpq_db7422df-8be7-4a79-9981-b5fd2348e4c4/extract/0.log" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.888731 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-77hn5_20bc6d13-f85d-423a-9782-097696fe3a0a/kube-rbac-proxy/0.log" Oct 08 15:48:11 crc kubenswrapper[4789]: I1008 15:48:11.953376 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-k8txt_78f0cb36-5bad-4021-b464-c5da9318404a/kube-rbac-proxy/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.010614 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-77hn5_20bc6d13-f85d-423a-9782-097696fe3a0a/manager/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.196318 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-k8txt_78f0cb36-5bad-4021-b464-c5da9318404a/manager/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.211631 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ptjbw_0aa40208-3985-4820-809f-ca1f94b05036/kube-rbac-proxy/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.247347 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ptjbw_0aa40208-3985-4820-809f-ca1f94b05036/manager/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.397701 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-tggcc_00c51d73-7d74-484a-8773-41cbb40c69f9/kube-rbac-proxy/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.443973 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-tggcc_00c51d73-7d74-484a-8773-41cbb40c69f9/manager/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.554283 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-vbtkd_0d8d4170-6f10-4d04-b443-297aada2dc10/kube-rbac-proxy/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.591125 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-vbtkd_0d8d4170-6f10-4d04-b443-297aada2dc10/manager/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.734854 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-w5bgc_f46571c0-32a7-4202-b308-54446adfe3d1/manager/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.740708 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a648d189-74dd-4364-b047-c9e535d0d626" path="/var/lib/kubelet/pods/a648d189-74dd-4364-b047-c9e535d0d626/volumes" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.742129 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-w5bgc_f46571c0-32a7-4202-b308-54446adfe3d1/kube-rbac-proxy/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.807805 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-wl7d2_e8242a68-738b-4fcd-abaa-22a18790ea28/kube-rbac-proxy/0.log" Oct 08 15:48:12 crc kubenswrapper[4789]: I1008 15:48:12.938366 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-zslsw_bff33916-174d-4fd4-a5e4-91f7cacb6e52/kube-rbac-proxy/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.001662 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-zslsw_bff33916-174d-4fd4-a5e4-91f7cacb6e52/manager/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.093288 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-wl7d2_e8242a68-738b-4fcd-abaa-22a18790ea28/manager/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.237691 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-bvjfl_f7b65ead-d9d5-4f06-80de-ae5c85dcbd45/kube-rbac-proxy/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.245922 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-bvjfl_f7b65ead-d9d5-4f06-80de-ae5c85dcbd45/manager/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.359792 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-pzg7c_8e51126b-8f41-4627-9f68-fb1255bda5bd/kube-rbac-proxy/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.425337 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-pzg7c_8e51126b-8f41-4627-9f68-fb1255bda5bd/manager/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.526495 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-7gsbz_3c0f54fe-d45f-44c2-9f09-4065c9621903/kube-rbac-proxy/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.590269 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-7gsbz_3c0f54fe-d45f-44c2-9f09-4065c9621903/manager/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.711652 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-2j5bd_c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4/kube-rbac-proxy/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.833918 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-2j5bd_c814d0bb-5b3a-48e8-a6c1-4ae2c759dec4/manager/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.900753 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-7trlm_c9e3e3c8-447f-4880-b54b-1d5601271216/kube-rbac-proxy/0.log" Oct 08 15:48:13 crc kubenswrapper[4789]: I1008 15:48:13.994501 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-7trlm_c9e3e3c8-447f-4880-b54b-1d5601271216/manager/0.log" Oct 08 15:48:14 crc kubenswrapper[4789]: I1008 15:48:14.052382 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-2kzff_4e7cd567-305f-417d-bac8-7bf4caf22693/kube-rbac-proxy/0.log" Oct 08 15:48:14 crc kubenswrapper[4789]: I1008 15:48:14.128024 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-2kzff_4e7cd567-305f-417d-bac8-7bf4caf22693/manager/0.log" Oct 08 15:48:14 crc kubenswrapper[4789]: I1008 15:48:14.275529 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h_6f893fd6-e5cc-42b9-ac46-77567818c0ef/kube-rbac-proxy/0.log" Oct 08 15:48:14 crc kubenswrapper[4789]: I1008 15:48:14.309922 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d45q9h_6f893fd6-e5cc-42b9-ac46-77567818c0ef/manager/0.log" Oct 08 15:48:14 crc kubenswrapper[4789]: I1008 15:48:14.448779 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-766b688d6c-zfshx_d174fba3-84a1-4de8-8b32-4f97d1880a78/kube-rbac-proxy/0.log" Oct 08 15:48:14 crc kubenswrapper[4789]: I1008 15:48:14.639004 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5cf8d5fc8d-4524r_9eaa64d7-4d19-436b-a954-ffd75b6d6cfd/kube-rbac-proxy/0.log" Oct 08 15:48:14 crc kubenswrapper[4789]: I1008 15:48:14.721348 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-swzf7_cb5f8cb4-1588-4ae9-88b1-890831151eb4/registry-server/0.log" Oct 08 15:48:14 crc kubenswrapper[4789]: I1008 15:48:14.792662 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5cf8d5fc8d-4524r_9eaa64d7-4d19-436b-a954-ffd75b6d6cfd/operator/0.log" Oct 08 15:48:14 crc kubenswrapper[4789]: I1008 15:48:14.857207 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-6f9x9_5578f70a-ad1b-47f8-a6ce-5dc726788ded/kube-rbac-proxy/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.036952 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-6f9x9_5578f70a-ad1b-47f8-a6ce-5dc726788ded/manager/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.098204 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-4xgrb_67d8c63c-cc55-4be0-bea4-8371ac8e333b/kube-rbac-proxy/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.108577 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-4xgrb_67d8c63c-cc55-4be0-bea4-8371ac8e333b/manager/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.279666 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-cpdl5_8bc03843-9303-451e-b933-c450aaad7401/operator/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.367712 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-nfllk_bc485c37-780a-4509-b3f3-aaf1714a9edb/kube-rbac-proxy/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.544103 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-775776c574-mgjfp_8efa82a4-efc9-4abe-b373-8828def5bb61/kube-rbac-proxy/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.555647 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-nfllk_bc485c37-780a-4509-b3f3-aaf1714a9edb/manager/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.795442 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-74665f6cdc-x52cc_08eb1945-dc5b-4f43-9fad-cd2b7e3102be/kube-rbac-proxy/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.800334 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-74665f6cdc-x52cc_08eb1945-dc5b-4f43-9fad-cd2b7e3102be/manager/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.904836 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-775776c574-mgjfp_8efa82a4-efc9-4abe-b373-8828def5bb61/manager/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.985214 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c7dbfbcbd-7xrj5_8bdbf6d1-353c-42ba-80df-e60d6b424f55/kube-rbac-proxy/0.log" Oct 08 15:48:15 crc kubenswrapper[4789]: I1008 15:48:15.991876 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-766b688d6c-zfshx_d174fba3-84a1-4de8-8b32-4f97d1880a78/manager/0.log" Oct 08 15:48:16 crc kubenswrapper[4789]: I1008 15:48:16.120145 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c7dbfbcbd-7xrj5_8bdbf6d1-353c-42ba-80df-e60d6b424f55/manager/0.log" Oct 08 15:48:26 crc kubenswrapper[4789]: I1008 15:48:26.433868 4789 patch_prober.go:28] interesting pod/machine-config-daemon-94gx2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 15:48:26 crc kubenswrapper[4789]: I1008 15:48:26.434430 4789 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 15:48:26 crc kubenswrapper[4789]: I1008 15:48:26.434479 4789 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" Oct 08 15:48:26 crc kubenswrapper[4789]: I1008 15:48:26.435288 4789 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a"} pod="openshift-machine-config-operator/machine-config-daemon-94gx2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 15:48:26 crc kubenswrapper[4789]: I1008 15:48:26.435334 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerName="machine-config-daemon" containerID="cri-o://dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" gracePeriod=600 Oct 08 15:48:26 crc kubenswrapper[4789]: E1008 15:48:26.568693 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:48:26 crc kubenswrapper[4789]: I1008 15:48:26.848295 4789 generic.go:334] "Generic (PLEG): container finished" podID="2b2e313a-834b-47fb-adae-992853bf6a6a" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" exitCode=0 Oct 08 15:48:26 crc kubenswrapper[4789]: I1008 15:48:26.848357 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerDied","Data":"dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a"} Oct 08 15:48:26 crc kubenswrapper[4789]: I1008 15:48:26.848642 4789 scope.go:117] "RemoveContainer" containerID="bd9daa530e0df0e138a935950ee0c210dd40007a2821f368084c00d54709c8aa" Oct 08 15:48:26 crc kubenswrapper[4789]: I1008 15:48:26.849550 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:48:26 crc kubenswrapper[4789]: E1008 15:48:26.849907 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:48:30 crc kubenswrapper[4789]: I1008 15:48:30.453467 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-c8rg8_cbd6f891-627b-4f0b-ac4d-40e30d01fb89/control-plane-machine-set-operator/0.log" Oct 08 15:48:30 crc kubenswrapper[4789]: I1008 15:48:30.645376 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wmqxc_bc2cc553-6ac7-44a6-948c-c1363836923d/machine-api-operator/0.log" Oct 08 15:48:30 crc kubenswrapper[4789]: I1008 15:48:30.660402 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wmqxc_bc2cc553-6ac7-44a6-948c-c1363836923d/kube-rbac-proxy/0.log" Oct 08 15:48:37 crc kubenswrapper[4789]: I1008 15:48:37.730157 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:48:37 crc kubenswrapper[4789]: E1008 15:48:37.731047 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:48:41 crc kubenswrapper[4789]: I1008 15:48:41.689183 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-qtnnc_a2f9dc0d-e779-42af-a82f-e3dc45ed7b6d/cert-manager-controller/0.log" Oct 08 15:48:41 crc kubenswrapper[4789]: I1008 15:48:41.835528 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-5vfnw_e05b78ba-5b41-4c35-b613-87ff25afd63b/cert-manager-cainjector/0.log" Oct 08 15:48:41 crc kubenswrapper[4789]: I1008 15:48:41.894104 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dkqb4_cf871183-8c76-4ab3-af84-a483ad387b7b/cert-manager-webhook/0.log" Oct 08 15:48:48 crc kubenswrapper[4789]: I1008 15:48:48.730590 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:48:48 crc kubenswrapper[4789]: E1008 15:48:48.731405 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:48:52 crc kubenswrapper[4789]: I1008 15:48:52.564465 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-4cxwh_b5615e4f-312c-49dc-9689-3662e0ad7ac8/nmstate-console-plugin/0.log" Oct 08 15:48:52 crc kubenswrapper[4789]: I1008 15:48:52.782374 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2qnx8_764cebf9-8453-4c3b-afe4-c6f56b39d87f/nmstate-handler/0.log" Oct 08 15:48:52 crc kubenswrapper[4789]: I1008 15:48:52.809030 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-l7cdj_e278e129-8d89-4095-a7f0-88a44cb19c97/kube-rbac-proxy/0.log" Oct 08 15:48:52 crc kubenswrapper[4789]: I1008 15:48:52.860515 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-l7cdj_e278e129-8d89-4095-a7f0-88a44cb19c97/nmstate-metrics/0.log" Oct 08 15:48:52 crc kubenswrapper[4789]: I1008 15:48:52.964549 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-cbqs2_333279f2-ccf6-44df-80ad-9097f320a3ac/nmstate-operator/0.log" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.039556 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-8ndgl_066fb44b-7e80-4cea-b5c9-ab9e208a9d52/nmstate-webhook/0.log" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.623780 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-85sxs"] Oct 08 15:48:53 crc kubenswrapper[4789]: E1008 15:48:53.624340 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a648d189-74dd-4364-b047-c9e535d0d626" containerName="container-00" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.624359 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="a648d189-74dd-4364-b047-c9e535d0d626" containerName="container-00" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.624618 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="a648d189-74dd-4364-b047-c9e535d0d626" containerName="container-00" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.626428 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.654139 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85sxs"] Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.746967 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-utilities\") pod \"redhat-operators-85sxs\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.747102 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-catalog-content\") pod \"redhat-operators-85sxs\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.747186 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42bcd\" (UniqueName: \"kubernetes.io/projected/babca2b4-63b8-4b92-8f42-2dd1179ff97a-kube-api-access-42bcd\") pod \"redhat-operators-85sxs\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.851356 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-catalog-content\") pod \"redhat-operators-85sxs\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.851429 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42bcd\" (UniqueName: \"kubernetes.io/projected/babca2b4-63b8-4b92-8f42-2dd1179ff97a-kube-api-access-42bcd\") pod \"redhat-operators-85sxs\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.851712 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-utilities\") pod \"redhat-operators-85sxs\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.851817 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-catalog-content\") pod \"redhat-operators-85sxs\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.852179 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-utilities\") pod \"redhat-operators-85sxs\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.877733 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42bcd\" (UniqueName: \"kubernetes.io/projected/babca2b4-63b8-4b92-8f42-2dd1179ff97a-kube-api-access-42bcd\") pod \"redhat-operators-85sxs\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:53 crc kubenswrapper[4789]: I1008 15:48:53.958179 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:48:54 crc kubenswrapper[4789]: I1008 15:48:54.527054 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85sxs"] Oct 08 15:48:55 crc kubenswrapper[4789]: I1008 15:48:55.118905 4789 generic.go:334] "Generic (PLEG): container finished" podID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerID="8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff" exitCode=0 Oct 08 15:48:55 crc kubenswrapper[4789]: I1008 15:48:55.118959 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sxs" event={"ID":"babca2b4-63b8-4b92-8f42-2dd1179ff97a","Type":"ContainerDied","Data":"8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff"} Oct 08 15:48:55 crc kubenswrapper[4789]: I1008 15:48:55.119219 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sxs" event={"ID":"babca2b4-63b8-4b92-8f42-2dd1179ff97a","Type":"ContainerStarted","Data":"ae616296ba0b4e94a2620f8c65eff00cd18d6a2eaaee27d0aa656cf553cda2e2"} Oct 08 15:48:57 crc kubenswrapper[4789]: I1008 15:48:57.140031 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sxs" event={"ID":"babca2b4-63b8-4b92-8f42-2dd1179ff97a","Type":"ContainerStarted","Data":"c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548"} Oct 08 15:49:01 crc kubenswrapper[4789]: I1008 15:49:01.730528 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:49:01 crc kubenswrapper[4789]: E1008 15:49:01.731295 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:49:02 crc kubenswrapper[4789]: I1008 15:49:02.212770 4789 generic.go:334] "Generic (PLEG): container finished" podID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerID="c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548" exitCode=0 Oct 08 15:49:02 crc kubenswrapper[4789]: I1008 15:49:02.212849 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sxs" event={"ID":"babca2b4-63b8-4b92-8f42-2dd1179ff97a","Type":"ContainerDied","Data":"c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548"} Oct 08 15:49:03 crc kubenswrapper[4789]: I1008 15:49:03.224976 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sxs" event={"ID":"babca2b4-63b8-4b92-8f42-2dd1179ff97a","Type":"ContainerStarted","Data":"a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b"} Oct 08 15:49:03 crc kubenswrapper[4789]: I1008 15:49:03.248392 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-85sxs" podStartSLOduration=2.597415198 podStartE2EDuration="10.24837103s" podCreationTimestamp="2025-10-08 15:48:53 +0000 UTC" firstStartedPulling="2025-10-08 15:48:55.120573892 +0000 UTC m=+6475.027321384" lastFinishedPulling="2025-10-08 15:49:02.771529724 +0000 UTC m=+6482.678277216" observedRunningTime="2025-10-08 15:49:03.243392134 +0000 UTC m=+6483.150139626" watchObservedRunningTime="2025-10-08 15:49:03.24837103 +0000 UTC m=+6483.155118512" Oct 08 15:49:03 crc kubenswrapper[4789]: I1008 15:49:03.959302 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:49:03 crc kubenswrapper[4789]: I1008 15:49:03.959758 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:49:05 crc kubenswrapper[4789]: I1008 15:49:05.015547 4789 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-85sxs" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerName="registry-server" probeResult="failure" output=< Oct 08 15:49:05 crc kubenswrapper[4789]: timeout: failed to connect service ":50051" within 1s Oct 08 15:49:05 crc kubenswrapper[4789]: > Oct 08 15:49:07 crc kubenswrapper[4789]: I1008 15:49:07.944597 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-ggzfl_9f942609-b5dc-4433-8767-a9bd36b185f9/kube-rbac-proxy/0.log" Oct 08 15:49:07 crc kubenswrapper[4789]: I1008 15:49:07.971875 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-ggzfl_9f942609-b5dc-4433-8767-a9bd36b185f9/controller/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.156680 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-frr-files/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.449714 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-metrics/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.452442 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-reloader/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.460037 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-frr-files/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.484644 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-reloader/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.657248 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-frr-files/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.694980 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-metrics/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.695066 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-reloader/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.740465 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-metrics/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.867172 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-metrics/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.867360 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-frr-files/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.877477 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/cp-reloader/0.log" Oct 08 15:49:08 crc kubenswrapper[4789]: I1008 15:49:08.934261 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/controller/0.log" Oct 08 15:49:09 crc kubenswrapper[4789]: I1008 15:49:09.068379 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/frr-metrics/0.log" Oct 08 15:49:09 crc kubenswrapper[4789]: I1008 15:49:09.073375 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/kube-rbac-proxy/0.log" Oct 08 15:49:09 crc kubenswrapper[4789]: I1008 15:49:09.122078 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/kube-rbac-proxy-frr/0.log" Oct 08 15:49:09 crc kubenswrapper[4789]: I1008 15:49:09.313685 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/reloader/0.log" Oct 08 15:49:09 crc kubenswrapper[4789]: I1008 15:49:09.387086 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-qx7cs_988fe042-8dd6-4398-af3d-ae201442f223/frr-k8s-webhook-server/0.log" Oct 08 15:49:09 crc kubenswrapper[4789]: I1008 15:49:09.622433 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-69d56bf45b-9wq7v_222ee176-36f5-4f7e-b6cf-8c526a0008d6/manager/0.log" Oct 08 15:49:09 crc kubenswrapper[4789]: I1008 15:49:09.847713 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-646mh_b76ec5f4-3801-4d1a-9b64-5b7beebc4100/kube-rbac-proxy/0.log" Oct 08 15:49:09 crc kubenswrapper[4789]: I1008 15:49:09.855658 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-655bc78868-jzwhv_067b91bf-82a6-4d6d-bf91-6f0cf31fa6dd/webhook-server/0.log" Oct 08 15:49:10 crc kubenswrapper[4789]: I1008 15:49:10.601180 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-646mh_b76ec5f4-3801-4d1a-9b64-5b7beebc4100/speaker/0.log" Oct 08 15:49:10 crc kubenswrapper[4789]: I1008 15:49:10.762320 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bq6zf_56b10d37-ef4c-4e34-87d1-2c15d04d63bc/frr/0.log" Oct 08 15:49:14 crc kubenswrapper[4789]: I1008 15:49:14.015576 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:49:14 crc kubenswrapper[4789]: I1008 15:49:14.086265 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:49:14 crc kubenswrapper[4789]: I1008 15:49:14.260557 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85sxs"] Oct 08 15:49:15 crc kubenswrapper[4789]: I1008 15:49:15.337579 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-85sxs" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerName="registry-server" containerID="cri-o://a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b" gracePeriod=2 Oct 08 15:49:15 crc kubenswrapper[4789]: I1008 15:49:15.840060 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.016063 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-catalog-content\") pod \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.016593 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-utilities\") pod \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.016857 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42bcd\" (UniqueName: \"kubernetes.io/projected/babca2b4-63b8-4b92-8f42-2dd1179ff97a-kube-api-access-42bcd\") pod \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\" (UID: \"babca2b4-63b8-4b92-8f42-2dd1179ff97a\") " Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.017315 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-utilities" (OuterVolumeSpecName: "utilities") pod "babca2b4-63b8-4b92-8f42-2dd1179ff97a" (UID: "babca2b4-63b8-4b92-8f42-2dd1179ff97a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.017861 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.024702 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/babca2b4-63b8-4b92-8f42-2dd1179ff97a-kube-api-access-42bcd" (OuterVolumeSpecName: "kube-api-access-42bcd") pod "babca2b4-63b8-4b92-8f42-2dd1179ff97a" (UID: "babca2b4-63b8-4b92-8f42-2dd1179ff97a"). InnerVolumeSpecName "kube-api-access-42bcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.117403 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "babca2b4-63b8-4b92-8f42-2dd1179ff97a" (UID: "babca2b4-63b8-4b92-8f42-2dd1179ff97a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.120650 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42bcd\" (UniqueName: \"kubernetes.io/projected/babca2b4-63b8-4b92-8f42-2dd1179ff97a-kube-api-access-42bcd\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.120716 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/babca2b4-63b8-4b92-8f42-2dd1179ff97a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.347696 4789 generic.go:334] "Generic (PLEG): container finished" podID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerID="a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b" exitCode=0 Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.347748 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sxs" event={"ID":"babca2b4-63b8-4b92-8f42-2dd1179ff97a","Type":"ContainerDied","Data":"a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b"} Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.347793 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sxs" event={"ID":"babca2b4-63b8-4b92-8f42-2dd1179ff97a","Type":"ContainerDied","Data":"ae616296ba0b4e94a2620f8c65eff00cd18d6a2eaaee27d0aa656cf553cda2e2"} Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.347817 4789 scope.go:117] "RemoveContainer" containerID="a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.347819 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85sxs" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.388714 4789 scope.go:117] "RemoveContainer" containerID="c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.389267 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85sxs"] Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.405637 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-85sxs"] Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.415641 4789 scope.go:117] "RemoveContainer" containerID="8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.476404 4789 scope.go:117] "RemoveContainer" containerID="a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b" Oct 08 15:49:16 crc kubenswrapper[4789]: E1008 15:49:16.476792 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b\": container with ID starting with a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b not found: ID does not exist" containerID="a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.476826 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b"} err="failed to get container status \"a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b\": rpc error: code = NotFound desc = could not find container \"a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b\": container with ID starting with a02a3ed51567525455ad965dc3bce0747f01d96a6a3f7728d40baf68f6c76a5b not found: ID does not exist" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.476855 4789 scope.go:117] "RemoveContainer" containerID="c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548" Oct 08 15:49:16 crc kubenswrapper[4789]: E1008 15:49:16.477244 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548\": container with ID starting with c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548 not found: ID does not exist" containerID="c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.477263 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548"} err="failed to get container status \"c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548\": rpc error: code = NotFound desc = could not find container \"c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548\": container with ID starting with c069d8a9bed21071836379c75b601e39651d54f58684d81d55251242a7c24548 not found: ID does not exist" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.477276 4789 scope.go:117] "RemoveContainer" containerID="8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff" Oct 08 15:49:16 crc kubenswrapper[4789]: E1008 15:49:16.477689 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff\": container with ID starting with 8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff not found: ID does not exist" containerID="8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.477789 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff"} err="failed to get container status \"8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff\": rpc error: code = NotFound desc = could not find container \"8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff\": container with ID starting with 8e746763ff870fb031b4fa230c3e3b8ab4c2841234f5b8ccf5e72b6626df1cff not found: ID does not exist" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.730370 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:49:16 crc kubenswrapper[4789]: E1008 15:49:16.730736 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:49:16 crc kubenswrapper[4789]: I1008 15:49:16.743829 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" path="/var/lib/kubelet/pods/babca2b4-63b8-4b92-8f42-2dd1179ff97a/volumes" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.071364 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/util/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.257597 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/util/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.294474 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/pull/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.328756 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/pull/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.536781 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/extract/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.584299 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/util/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.584924 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rjvm2_5084833a-0245-4282-8816-9d722342b260/pull/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.756367 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/util/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.899547 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/pull/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.903610 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/util/0.log" Oct 08 15:49:23 crc kubenswrapper[4789]: I1008 15:49:23.933076 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/pull/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.155964 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/extract/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.210319 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/pull/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.254450 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dcf7hn_fa1275e9-3700-47b0-b09f-a68a046cc1e4/util/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.378045 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-utilities/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.569430 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-utilities/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.608478 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-content/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.639698 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-content/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.757079 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-utilities/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.780573 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/extract-content/0.log" Oct 08 15:49:24 crc kubenswrapper[4789]: I1008 15:49:24.979582 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-utilities/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.051583 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-f96jk_6bdc52e6-9e72-411a-9558-0607f78fe55c/registry-server/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.113264 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-utilities/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.181537 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-content/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.212556 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-content/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.392511 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-utilities/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.414228 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/extract-content/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.615557 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/util/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.685258 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qztcr_110bf958-13b6-4727-9d9b-ff08e0ac9146/registry-server/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.843583 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/util/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.858093 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/pull/0.log" Oct 08 15:49:25 crc kubenswrapper[4789]: I1008 15:49:25.864710 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/pull/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.021342 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/util/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.078349 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/pull/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.083641 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ccf5wc_b98286a8-216e-4507-9353-f983186c2e7b/extract/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.191381 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4lxzv_c59c8d39-3840-48cf-b510-4e07d9293fc9/marketplace-operator/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.262798 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-utilities/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.441102 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-utilities/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.466244 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-content/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.487443 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-content/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.617755 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-content/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.640476 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/extract-utilities/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.667251 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-utilities/0.log" Oct 08 15:49:26 crc kubenswrapper[4789]: I1008 15:49:26.854498 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gxpd9_d530161d-7c51-48bb-8c08-0f48a5842e3e/registry-server/0.log" Oct 08 15:49:27 crc kubenswrapper[4789]: I1008 15:49:27.145293 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-content/0.log" Oct 08 15:49:27 crc kubenswrapper[4789]: I1008 15:49:27.147927 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-utilities/0.log" Oct 08 15:49:27 crc kubenswrapper[4789]: I1008 15:49:27.177184 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-content/0.log" Oct 08 15:49:27 crc kubenswrapper[4789]: I1008 15:49:27.287007 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-utilities/0.log" Oct 08 15:49:27 crc kubenswrapper[4789]: I1008 15:49:27.330730 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/extract-content/0.log" Oct 08 15:49:27 crc kubenswrapper[4789]: I1008 15:49:27.729672 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:49:27 crc kubenswrapper[4789]: E1008 15:49:27.730083 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:49:28 crc kubenswrapper[4789]: I1008 15:49:28.090921 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pqnwd_51eae247-9044-49e7-a270-6a943c12a3ac/registry-server/0.log" Oct 08 15:49:39 crc kubenswrapper[4789]: I1008 15:49:39.807553 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-lwgzt_5f7cb3da-f7e4-49ec-9ad4-dea3650cf5ec/prometheus-operator/0.log" Oct 08 15:49:39 crc kubenswrapper[4789]: I1008 15:49:39.922562 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-c5b6ff9c9-7987l_9451e6d9-fa5b-4209-a332-a65412994d55/prometheus-operator-admission-webhook/0.log" Oct 08 15:49:40 crc kubenswrapper[4789]: I1008 15:49:40.005634 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-c5b6ff9c9-srkvh_7141b3e4-ea4a-4476-84d2-57670fe5e932/prometheus-operator-admission-webhook/0.log" Oct 08 15:49:40 crc kubenswrapper[4789]: I1008 15:49:40.117343 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-tclgl_643a32e3-75dd-44f5-bbac-e2d917641a29/operator/0.log" Oct 08 15:49:40 crc kubenswrapper[4789]: I1008 15:49:40.180141 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-rbwl7_1cc6e22a-f412-459c-a8b4-5ee1fabe0bd5/perses-operator/0.log" Oct 08 15:49:41 crc kubenswrapper[4789]: I1008 15:49:41.730274 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:49:41 crc kubenswrapper[4789]: E1008 15:49:41.730810 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:49:54 crc kubenswrapper[4789]: I1008 15:49:54.730982 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:49:54 crc kubenswrapper[4789]: E1008 15:49:54.731797 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:50:02 crc kubenswrapper[4789]: E1008 15:50:02.066697 4789 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.214:57090->38.102.83.214:38325: read tcp 38.102.83.214:57090->38.102.83.214:38325: read: connection reset by peer Oct 08 15:50:05 crc kubenswrapper[4789]: I1008 15:50:05.730550 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:50:05 crc kubenswrapper[4789]: E1008 15:50:05.731208 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:50:17 crc kubenswrapper[4789]: I1008 15:50:17.730512 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:50:17 crc kubenswrapper[4789]: E1008 15:50:17.731597 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:50:32 crc kubenswrapper[4789]: I1008 15:50:32.731145 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:50:32 crc kubenswrapper[4789]: E1008 15:50:32.732030 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:50:45 crc kubenswrapper[4789]: I1008 15:50:45.730610 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:50:45 crc kubenswrapper[4789]: E1008 15:50:45.731549 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:50:56 crc kubenswrapper[4789]: I1008 15:50:56.731144 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:50:56 crc kubenswrapper[4789]: E1008 15:50:56.732591 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:51:10 crc kubenswrapper[4789]: I1008 15:51:10.741467 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:51:10 crc kubenswrapper[4789]: E1008 15:51:10.742352 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:51:22 crc kubenswrapper[4789]: I1008 15:51:22.731405 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:51:22 crc kubenswrapper[4789]: E1008 15:51:22.732454 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:51:33 crc kubenswrapper[4789]: I1008 15:51:33.730243 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:51:33 crc kubenswrapper[4789]: E1008 15:51:33.731236 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.008084 4789 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jhzz8"] Oct 08 15:51:39 crc kubenswrapper[4789]: E1008 15:51:39.009278 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerName="extract-utilities" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.009296 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerName="extract-utilities" Oct 08 15:51:39 crc kubenswrapper[4789]: E1008 15:51:39.009497 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerName="registry-server" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.009510 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerName="registry-server" Oct 08 15:51:39 crc kubenswrapper[4789]: E1008 15:51:39.009536 4789 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerName="extract-content" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.009544 4789 state_mem.go:107] "Deleted CPUSet assignment" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerName="extract-content" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.009789 4789 memory_manager.go:354] "RemoveStaleState removing state" podUID="babca2b4-63b8-4b92-8f42-2dd1179ff97a" containerName="registry-server" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.015689 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.025954 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jhzz8"] Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.196821 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-utilities\") pod \"redhat-marketplace-jhzz8\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.197198 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg8ln\" (UniqueName: \"kubernetes.io/projected/73b1a666-0f48-4af4-bd53-61eac893f857-kube-api-access-fg8ln\") pod \"redhat-marketplace-jhzz8\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.197303 4789 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-catalog-content\") pod \"redhat-marketplace-jhzz8\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.299869 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-utilities\") pod \"redhat-marketplace-jhzz8\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.299927 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg8ln\" (UniqueName: \"kubernetes.io/projected/73b1a666-0f48-4af4-bd53-61eac893f857-kube-api-access-fg8ln\") pod \"redhat-marketplace-jhzz8\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.299972 4789 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-catalog-content\") pod \"redhat-marketplace-jhzz8\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.300398 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-utilities\") pod \"redhat-marketplace-jhzz8\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.300469 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-catalog-content\") pod \"redhat-marketplace-jhzz8\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.326594 4789 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg8ln\" (UniqueName: \"kubernetes.io/projected/73b1a666-0f48-4af4-bd53-61eac893f857-kube-api-access-fg8ln\") pod \"redhat-marketplace-jhzz8\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.343551 4789 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:39 crc kubenswrapper[4789]: I1008 15:51:39.852560 4789 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jhzz8"] Oct 08 15:51:40 crc kubenswrapper[4789]: I1008 15:51:40.763519 4789 generic.go:334] "Generic (PLEG): container finished" podID="73b1a666-0f48-4af4-bd53-61eac893f857" containerID="3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4" exitCode=0 Oct 08 15:51:40 crc kubenswrapper[4789]: I1008 15:51:40.763807 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jhzz8" event={"ID":"73b1a666-0f48-4af4-bd53-61eac893f857","Type":"ContainerDied","Data":"3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4"} Oct 08 15:51:40 crc kubenswrapper[4789]: I1008 15:51:40.763826 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jhzz8" event={"ID":"73b1a666-0f48-4af4-bd53-61eac893f857","Type":"ContainerStarted","Data":"3b3fb4e2c44b658360e37513bece1ffcb55802ad20504867be5b574a7517c55f"} Oct 08 15:51:40 crc kubenswrapper[4789]: I1008 15:51:40.765862 4789 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 15:51:42 crc kubenswrapper[4789]: I1008 15:51:42.782772 4789 generic.go:334] "Generic (PLEG): container finished" podID="73b1a666-0f48-4af4-bd53-61eac893f857" containerID="37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910" exitCode=0 Oct 08 15:51:42 crc kubenswrapper[4789]: I1008 15:51:42.782874 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jhzz8" event={"ID":"73b1a666-0f48-4af4-bd53-61eac893f857","Type":"ContainerDied","Data":"37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910"} Oct 08 15:51:43 crc kubenswrapper[4789]: I1008 15:51:43.798280 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jhzz8" event={"ID":"73b1a666-0f48-4af4-bd53-61eac893f857","Type":"ContainerStarted","Data":"9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf"} Oct 08 15:51:43 crc kubenswrapper[4789]: I1008 15:51:43.828842 4789 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jhzz8" podStartSLOduration=3.290879051 podStartE2EDuration="5.82882343s" podCreationTimestamp="2025-10-08 15:51:38 +0000 UTC" firstStartedPulling="2025-10-08 15:51:40.765468023 +0000 UTC m=+6640.672215515" lastFinishedPulling="2025-10-08 15:51:43.303412402 +0000 UTC m=+6643.210159894" observedRunningTime="2025-10-08 15:51:43.820260927 +0000 UTC m=+6643.727008429" watchObservedRunningTime="2025-10-08 15:51:43.82882343 +0000 UTC m=+6643.735570922" Oct 08 15:51:48 crc kubenswrapper[4789]: I1008 15:51:48.734496 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:51:48 crc kubenswrapper[4789]: E1008 15:51:48.735422 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:51:49 crc kubenswrapper[4789]: I1008 15:51:49.344144 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:49 crc kubenswrapper[4789]: I1008 15:51:49.344317 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:49 crc kubenswrapper[4789]: I1008 15:51:49.400932 4789 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:49 crc kubenswrapper[4789]: I1008 15:51:49.898960 4789 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:49 crc kubenswrapper[4789]: I1008 15:51:49.952148 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jhzz8"] Oct 08 15:51:51 crc kubenswrapper[4789]: I1008 15:51:51.868498 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jhzz8" podUID="73b1a666-0f48-4af4-bd53-61eac893f857" containerName="registry-server" containerID="cri-o://9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf" gracePeriod=2 Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.377511 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.417826 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-utilities\") pod \"73b1a666-0f48-4af4-bd53-61eac893f857\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.418154 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-catalog-content\") pod \"73b1a666-0f48-4af4-bd53-61eac893f857\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.418226 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg8ln\" (UniqueName: \"kubernetes.io/projected/73b1a666-0f48-4af4-bd53-61eac893f857-kube-api-access-fg8ln\") pod \"73b1a666-0f48-4af4-bd53-61eac893f857\" (UID: \"73b1a666-0f48-4af4-bd53-61eac893f857\") " Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.418935 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-utilities" (OuterVolumeSpecName: "utilities") pod "73b1a666-0f48-4af4-bd53-61eac893f857" (UID: "73b1a666-0f48-4af4-bd53-61eac893f857"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.431091 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73b1a666-0f48-4af4-bd53-61eac893f857" (UID: "73b1a666-0f48-4af4-bd53-61eac893f857"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.433672 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b1a666-0f48-4af4-bd53-61eac893f857-kube-api-access-fg8ln" (OuterVolumeSpecName: "kube-api-access-fg8ln") pod "73b1a666-0f48-4af4-bd53-61eac893f857" (UID: "73b1a666-0f48-4af4-bd53-61eac893f857"). InnerVolumeSpecName "kube-api-access-fg8ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.520078 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg8ln\" (UniqueName: \"kubernetes.io/projected/73b1a666-0f48-4af4-bd53-61eac893f857-kube-api-access-fg8ln\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.520123 4789 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.520136 4789 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b1a666-0f48-4af4-bd53-61eac893f857-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.880657 4789 generic.go:334] "Generic (PLEG): container finished" podID="73b1a666-0f48-4af4-bd53-61eac893f857" containerID="9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf" exitCode=0 Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.880696 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jhzz8" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.880713 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jhzz8" event={"ID":"73b1a666-0f48-4af4-bd53-61eac893f857","Type":"ContainerDied","Data":"9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf"} Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.882275 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jhzz8" event={"ID":"73b1a666-0f48-4af4-bd53-61eac893f857","Type":"ContainerDied","Data":"3b3fb4e2c44b658360e37513bece1ffcb55802ad20504867be5b574a7517c55f"} Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.882294 4789 scope.go:117] "RemoveContainer" containerID="9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.911135 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jhzz8"] Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.922559 4789 scope.go:117] "RemoveContainer" containerID="37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.936653 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jhzz8"] Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.946266 4789 scope.go:117] "RemoveContainer" containerID="3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.993159 4789 scope.go:117] "RemoveContainer" containerID="9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf" Oct 08 15:51:52 crc kubenswrapper[4789]: E1008 15:51:52.995288 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf\": container with ID starting with 9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf not found: ID does not exist" containerID="9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.995530 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf"} err="failed to get container status \"9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf\": rpc error: code = NotFound desc = could not find container \"9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf\": container with ID starting with 9354e981cbb70ec60b474469c86cc165e4045cac26d15b599d6986370121beaf not found: ID does not exist" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.995670 4789 scope.go:117] "RemoveContainer" containerID="37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910" Oct 08 15:51:52 crc kubenswrapper[4789]: E1008 15:51:52.996097 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910\": container with ID starting with 37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910 not found: ID does not exist" containerID="37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.996256 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910"} err="failed to get container status \"37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910\": rpc error: code = NotFound desc = could not find container \"37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910\": container with ID starting with 37ad8d49045384d5800b0926c78976198428167c9c324df2139bf05bc866e910 not found: ID does not exist" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.996338 4789 scope.go:117] "RemoveContainer" containerID="3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4" Oct 08 15:51:52 crc kubenswrapper[4789]: E1008 15:51:52.996606 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4\": container with ID starting with 3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4 not found: ID does not exist" containerID="3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4" Oct 08 15:51:52 crc kubenswrapper[4789]: I1008 15:51:52.996700 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4"} err="failed to get container status \"3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4\": rpc error: code = NotFound desc = could not find container \"3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4\": container with ID starting with 3869f6e2312b378fbda0991b47e6bb7d12649f12494530c82384f686e6cf85b4 not found: ID does not exist" Oct 08 15:51:54 crc kubenswrapper[4789]: I1008 15:51:54.744073 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b1a666-0f48-4af4-bd53-61eac893f857" path="/var/lib/kubelet/pods/73b1a666-0f48-4af4-bd53-61eac893f857/volumes" Oct 08 15:51:59 crc kubenswrapper[4789]: I1008 15:51:59.958051 4789 generic.go:334] "Generic (PLEG): container finished" podID="3401a218-11d5-4000-b639-1f2dccc3b80b" containerID="b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae" exitCode=0 Oct 08 15:51:59 crc kubenswrapper[4789]: I1008 15:51:59.958128 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ws6vd/must-gather-v68cr" event={"ID":"3401a218-11d5-4000-b639-1f2dccc3b80b","Type":"ContainerDied","Data":"b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae"} Oct 08 15:51:59 crc kubenswrapper[4789]: I1008 15:51:59.959318 4789 scope.go:117] "RemoveContainer" containerID="b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae" Oct 08 15:52:00 crc kubenswrapper[4789]: I1008 15:52:00.178181 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ws6vd_must-gather-v68cr_3401a218-11d5-4000-b639-1f2dccc3b80b/gather/0.log" Oct 08 15:52:03 crc kubenswrapper[4789]: I1008 15:52:03.729954 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:52:03 crc kubenswrapper[4789]: E1008 15:52:03.730752 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.151859 4789 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ws6vd/must-gather-v68cr"] Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.153396 4789 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ws6vd/must-gather-v68cr" podUID="3401a218-11d5-4000-b639-1f2dccc3b80b" containerName="copy" containerID="cri-o://272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151" gracePeriod=2 Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.164905 4789 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ws6vd/must-gather-v68cr"] Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.615463 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ws6vd_must-gather-v68cr_3401a218-11d5-4000-b639-1f2dccc3b80b/copy/0.log" Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.616143 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.704865 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3401a218-11d5-4000-b639-1f2dccc3b80b-must-gather-output\") pod \"3401a218-11d5-4000-b639-1f2dccc3b80b\" (UID: \"3401a218-11d5-4000-b639-1f2dccc3b80b\") " Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.705185 4789 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2s2w\" (UniqueName: \"kubernetes.io/projected/3401a218-11d5-4000-b639-1f2dccc3b80b-kube-api-access-h2s2w\") pod \"3401a218-11d5-4000-b639-1f2dccc3b80b\" (UID: \"3401a218-11d5-4000-b639-1f2dccc3b80b\") " Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.710710 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3401a218-11d5-4000-b639-1f2dccc3b80b-kube-api-access-h2s2w" (OuterVolumeSpecName: "kube-api-access-h2s2w") pod "3401a218-11d5-4000-b639-1f2dccc3b80b" (UID: "3401a218-11d5-4000-b639-1f2dccc3b80b"). InnerVolumeSpecName "kube-api-access-h2s2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.731189 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:52:14 crc kubenswrapper[4789]: E1008 15:52:14.731493 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.807848 4789 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2s2w\" (UniqueName: \"kubernetes.io/projected/3401a218-11d5-4000-b639-1f2dccc3b80b-kube-api-access-h2s2w\") on node \"crc\" DevicePath \"\"" Oct 08 15:52:14 crc kubenswrapper[4789]: I1008 15:52:14.915785 4789 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3401a218-11d5-4000-b639-1f2dccc3b80b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3401a218-11d5-4000-b639-1f2dccc3b80b" (UID: "3401a218-11d5-4000-b639-1f2dccc3b80b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.012548 4789 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3401a218-11d5-4000-b639-1f2dccc3b80b-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.135803 4789 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ws6vd_must-gather-v68cr_3401a218-11d5-4000-b639-1f2dccc3b80b/copy/0.log" Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.136315 4789 generic.go:334] "Generic (PLEG): container finished" podID="3401a218-11d5-4000-b639-1f2dccc3b80b" containerID="272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151" exitCode=143 Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.136370 4789 scope.go:117] "RemoveContainer" containerID="272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151" Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.136405 4789 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ws6vd/must-gather-v68cr" Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.168826 4789 scope.go:117] "RemoveContainer" containerID="b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae" Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.259698 4789 scope.go:117] "RemoveContainer" containerID="272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151" Oct 08 15:52:15 crc kubenswrapper[4789]: E1008 15:52:15.260188 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151\": container with ID starting with 272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151 not found: ID does not exist" containerID="272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151" Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.260220 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151"} err="failed to get container status \"272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151\": rpc error: code = NotFound desc = could not find container \"272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151\": container with ID starting with 272c58d5c9a8e5c9caa49db1efee8570ee677628a80161c4649b6ea4297f4151 not found: ID does not exist" Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.260255 4789 scope.go:117] "RemoveContainer" containerID="b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae" Oct 08 15:52:15 crc kubenswrapper[4789]: E1008 15:52:15.260675 4789 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae\": container with ID starting with b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae not found: ID does not exist" containerID="b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae" Oct 08 15:52:15 crc kubenswrapper[4789]: I1008 15:52:15.260696 4789 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae"} err="failed to get container status \"b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae\": rpc error: code = NotFound desc = could not find container \"b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae\": container with ID starting with b5cb008666a8a0e9bb125e718e732d95262416a9bd7d5e5862473d0b202dd1ae not found: ID does not exist" Oct 08 15:52:16 crc kubenswrapper[4789]: I1008 15:52:16.749342 4789 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3401a218-11d5-4000-b639-1f2dccc3b80b" path="/var/lib/kubelet/pods/3401a218-11d5-4000-b639-1f2dccc3b80b/volumes" Oct 08 15:52:25 crc kubenswrapper[4789]: I1008 15:52:25.732585 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:52:25 crc kubenswrapper[4789]: E1008 15:52:25.734819 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:52:27 crc kubenswrapper[4789]: I1008 15:52:27.027139 4789 scope.go:117] "RemoveContainer" containerID="1a22986e1795292d8b8e81e68095dc7d01d96a36733c8339a563f054315a04f3" Oct 08 15:52:36 crc kubenswrapper[4789]: I1008 15:52:36.730170 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:52:36 crc kubenswrapper[4789]: E1008 15:52:36.730915 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:52:50 crc kubenswrapper[4789]: I1008 15:52:50.736528 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:52:50 crc kubenswrapper[4789]: E1008 15:52:50.737314 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:53:01 crc kubenswrapper[4789]: I1008 15:53:01.730027 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:53:01 crc kubenswrapper[4789]: E1008 15:53:01.730736 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:53:14 crc kubenswrapper[4789]: I1008 15:53:14.730493 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:53:14 crc kubenswrapper[4789]: E1008 15:53:14.731585 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:53:25 crc kubenswrapper[4789]: I1008 15:53:25.730258 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:53:25 crc kubenswrapper[4789]: E1008 15:53:25.731128 4789 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-94gx2_openshift-machine-config-operator(2b2e313a-834b-47fb-adae-992853bf6a6a)\"" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" podUID="2b2e313a-834b-47fb-adae-992853bf6a6a" Oct 08 15:53:39 crc kubenswrapper[4789]: I1008 15:53:39.730675 4789 scope.go:117] "RemoveContainer" containerID="dbaae3bab97efa635af7e569827085e1929576f688f2de28497015f34fff871a" Oct 08 15:53:40 crc kubenswrapper[4789]: I1008 15:53:40.996067 4789 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-94gx2" event={"ID":"2b2e313a-834b-47fb-adae-992853bf6a6a","Type":"ContainerStarted","Data":"303d721c5231135d35cec522d27f7fb099670225c77226315dd2a44882584622"} Oct 08 15:54:27 crc kubenswrapper[4789]: I1008 15:54:27.132385 4789 scope.go:117] "RemoveContainer" containerID="388f57656d77819dcb55d723c361e9a49b2469e35dbdc2157e2de9cf48d8af5d" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071504500024442 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071504500017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071466664016524 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071466665015475 5ustar corecore